Matt's Blog

AI chat app and model distinction

We’re a solid three-plus years into AI chat apps. ChatGPT, Claude, Gemini, Grok, Perplexity, etc. From afar, they all look pretty similar, but each creates a distinctly different user experience.

One thing I found important early on when working with Large Language Models (LLMs) is recognizing that there’s a LOT more going on than just the underlying model. For any of these flagship AI chat apps, you’re never engaging directly with the foundational model. The model matters, for sure, but not as much as the application wrapped around it.

Putting a finer point on it, there are two ingredients that create differentiated experiences in these chat apps:

A great model embedded in a so-so application will be at best only so-so. A great app, even with a less capable foundation model, can still be amazing. The best of the best—which is hard to achieve—is having a great model and a great app at the same time.

In my day job, I work with organizations on AI projects, and this distinction shapes how I think about building with AI.

What all does the app do?

It’s hard to have a great app and model at the same time

Foundation models don’t stay the most powerful for long. There’s huge focus among model developers to continue innovating, and we see leapfrogging every month or so.

Meanwhile, the apps continually need to adapt to new capabilities that models support, as well as the nuances of communicating with different models.

Isn’t this all obvious? Why are you writing about it?

In short, no—it’s not obvious. Even experts in the AI space rarely speak about this distinction. I’ve been frustrated with this since the beginning of ChatGPT, and it’s continued to bug me for three years now.

Recently, Lex Fridman had a four hour episode (#490) discussing AI, starting with chat apps. They spent significant time discussing these applications, yet the conversation never quite touched on this fundamental distinction between the model and the application layer.

They know their stuff—I’m not suggesting they don’t “get it,” because they do. That said, I was hoping to hear some of the conversation delve into this distinction and the larger implications for AI projects outside of flagship chat apps.

There’s more nuance to it. For me, a big part of ChatGPT’s success has been the application experience they create. They’ve been great at it, had a huge head start, and still maintain a lead there.

But for people and organizations who aren’t also creating their own models, the focus should be on creating a great application—one that’s ready to swap in other models, which are increasingly becoming commodities.

This is where I like Microsoft’s capabilities to support developers

Note: I work at Microsoft. This is my opinion and I’m not speaking for Microsoft.

Microsoft Foundry is geared toward helping developers build intelligent applications. Despite some naming challenges (what it’s called isn’t the important thing), it brings significant capabilities relevant to this conversation:

Final thoughts

A few things to keep in mind as you go about your merry way:

#Ai #Llm #Chatgpt #Claude #Microsoft-Foundry #Application-Development #Product-Thinking