AIModels.fyi

AIModels.fyi

Losers (and winners) from the GPT-4o announcement

Some products are more toast than others

aimodels-fyi's avatar
aimodels-fyi
May 14, 2024
∙ Paid
“Hey ChatGPT, do I look lonely?”

Normally, I cover research papers, tools, and models for you. But today, I want to put my product hat on and give some ideas on who I think the biggest winners and losers from yesterday’s GPT-4o announcement. In case you missed it, this latest OpenAI model processes and generates text, audio, and images in real time. This "omni" model is a major step forward in creating more natural human-computer interactions.

Here's a look at who I think stands to gain and who might lose out from this development.

AIModels.fyi is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.

Context

GPT-4o, with "o" standing for "omni," handles text, audio, and images simultaneously. It responds to audio inputs in just 232 milliseconds, matching human conversational speeds. It's faster, more accurate, and 50% cheaper than its predecessors. The model excels in non-English languages and multimodal tasks, making it a game-changer in the AI arms race.

By integrating text, audio, and visual processing into a single system (and, from what I can tell, doing the inference directly from the inputs as opposed to first transcribing everything), GPT-4o sets a new standard for conversational AI. This model interprets and responds to inputs in real time, making interactions with AI more seamless and natural.

So, I think this is the key idea when looking at which other products stand to benefit or suffer from this release. First up, the losers…

User's avatar

Continue reading this post for free, courtesy of aimodels-fyi.

Or purchase a paid subscription.
© 2026 AIModels.fyi · Privacy ∙ Terms ∙ Collection notice
Start your SubstackGet the app
Substack is the home for great culture