GPT-4o

OpenAI's new flagship model that can reason across audio, vision, and text in real time.

Homepage

Hello GPT-4o

FAQ

1. What is GPT-4o and what makes it unique?

GPT-4o (‘o’ for ‘omni’) is a model that accepts any combination of text, audio, image, and video inputs and generates corresponding outputs in text, audio, and image formats. It excels in vision and audio understanding compared to existing models.

2. How fast is GPT-4o in responding to audio inputs?

GPT-4o can respond to audio inputs in as little as 232 milliseconds, with an average response time of 320 milliseconds, similar to human conversation response times.

3. How does GPT-4o compare to previous models in terms of performance and cost?

GPT-4o matches GPT-4 Turbo performance on text in English and code, with significant improvements in non-English languages. It is faster and 50% cheaper in the API compared to previous models.

4. What is the difference between GPT-4o and Voice Mode in ChatGPT?

Voice Mode in ChatGPT had latencies of 2.8 seconds (GPT-3.5) and 5.4 seconds (GPT-4) on average. With GPT-4o, a single new model is trained end-to-end across text, vision, and audio, providing more accurate and comprehensive responses.

5. How does GPT-4o perform in terms of reasoning intelligence and multilingual capabilities?

GPT-4o sets new high scores in reasoning intelligence and multilingual capabilities, achieving 88.7% on 0-shot general knowledge questions and 87.2% on traditional 5-shot tasks.

6. How does GPT-4o improve audio performance compared to previous models?

GPT-4o dramatically improves speech recognition and translation performance across all languages, particularly for lower-resourced languages.

7. How does GPT-4o perform on vision understanding evaluations?

GPT-4o achieves state-of-the-art performance on visual perception benchmarks, showcasing its strength in understanding and processing visual information.

8. What is the M3Exam benchmark, and how does GPT-4o perform on it?

M3Exam is a multilingual and vision evaluation benchmark, where GPT-4o outperforms GPT-4 across all languages, demonstrating its proficiency in handling diverse tasks.

9. Which languages have seen significant token reduction with GPT-4o's new tokenizer?

GPT-4o's new tokenizer has reduced the number of tokens significantly in languages such as Gujarati, Telugu, Tamil, Marathi, Hindi, Urdu, Arabic, Russian, Chinese, French, and English.

10. Is safety a priority in GPT-4o?

Yes, GPT-4o has safety features built-in to ensure responsible and ethical usage of the model.

Was this page helpful?