OpenAI has made a groundbreaking announcement with the launch of GPT-4o, its latest artificial intelligence model that boasts native support for reasoning across audio, visual, and text. The “o” in GPT-4o stands for “Omni”, highlighting its superior ability to understand and interpret various forms of data compared to its predecessor.
This new model represents a significant step towards more natural human-computer interaction, as it can seamlessly process combinations of text, audio, and images to generate responses. With a response time of 232 milliseconds for audio inputs, GPT-4o rivals human conversational speed.
Compared to the existing GPT-4 Turbo model, GPT-4o excels in audio understanding while matching performance in English text comprehension and coding. Additionally, it offers substantial improvements in processing non-English languages and understanding images.
OpenAI has also introduced the ChatGPT application for Apple’s macOS-based desktops, providing users with a convenient platform to interact with the GPT-4o model. The Voice Mode feature in ChatGPT has been enhanced with the new model, offering a more seamless conversational experience with reduced latency.
While the GPT-4o model is available to free tier users on ChatGPT, certain advanced features such as Voice Mode remain exclusive to paid subscribers. However, free users can now access features previously limited to paid tiers, including file uploads, memory capabilities, and browsing the GPT Store for custom bots.
Overall, the introduction of GPT-4o and its integration into ChatGPT represents a significant advancement in artificial intelligence technology, paving the way for more sophisticated and intuitive human-computer interactions.