Hello GPT-4o: OpenAI’s New Flagship Model

What Is The News About?

OpenAI is yet here again with an awesome model. GPT-4o (“o” for “omni”) is a step towards much more natural human-computer interaction—it accepts as input any combination of text, audio, image, and video and generates any combination of text, audio, and image outputs. It can respond to audio inputs in as little as 232 milliseconds, with an average of 320 milliseconds, which is similar to human response time(opens in a new window) in a conversation. It matches GPT-4 Turbo performance on text in English and code, with significant improvement on text in non-English languages, while also being much faster and 50% cheaper in the API. GPT-4o is especially better at vision and audio understanding compared to existing models.

Read: Recap of Day 1 Qualtrics X4 Summit: New Qualtrics AI Capabilities Will Fix Broken Digital Experiences

Why Is This News Important?

Meta started classifying photographs made by OpenAI, Midjourney, and other AI solutions in February, as the company stated. In a statement released last week, social media platform TikTok also announced plans to begin tagging similar photographs. At the same time as OpenAI has seen a flurry of new leadership, they have released a number of significant products in the past several years. Suddenly, in November 2023, Altman resigned as CEO of OpenAI. Just four days subsequent to a staff mutiny and a public apology from a board member of the corporation, Altman was rehired for the role. One of the requirements of the decision was that OpenAI reorganize its board of directors. Microsoft CEO Satya Nadella, whose business invested $10 billion in OpenAI last year, seems to have involvement for Altman’s return.


  1. Enhanced Interaction: GPT-4o enables seamless human-computer interaction across text, audio, image, and video inputs, fostering natural communication experiences.
  2. Improved Performance and Affordability: With response times comparable to human conversations and cost-effectiveness, GPT-4o outperforms previous models, benefiting both users and developers.
  3. Advanced Capabilities: GPT-4o excels in vision and audio understanding, providing superior performance in interpreting and generating content compared to existing models.


OpenAI unveiled the most recent version of the AI language model that powers their conversation bot ChatGPT and also revealed a desktop version of the bot. At a corporate event that was livestreamed, OpenAI Chief Technology Officer Mira Murati noted that the new model is faster and has more interactive capabilities than the previous model, GPT-4. According to Murati, the most recent version of GPT, codenamed GPT-4o, removes lag time from ChatGPT’s responses and lets users enter a new query to modify the bot’s response. According to Murati, the enhanced capability operates in reaction to visual, auditory, and textual cues.A different demonstration showed the device observing a researcher’s attempt to solve a mathematical equation using the camera on his phone and offering real-time assistance.

The distribution of GPT-4o is scheduled for the coming weeks, according to Murati. The company plans to deploy the product gradually to avoid abuse, he added. The most recent version of Open-AI’s AI language model, GPT-4, was released in March 2023. Sam Altman, CEO of OpenAI, informed ABC News a few days after GPT-4’s release that the product achieved a 90th percentile on the Uniform Bar Exam. It is also quite adept in writing code in the majority of programming languages and achieved an almost flawless score on the SAT Math exam.

Read: Decoding the “X” Factor in Healthcare Industry: Qualtrics’ Dr. Adrienne Boissy Explains the Science Behind Experience Management (XM)

[To share your insights with us as part of editorial or sponsored content, please write to [email protected]]

The post Hello GPT-4o: OpenAI’s New Flagship Model appeared first on AiThority.

Your custom text © Copyright 2024. All rights reserved.