OpenAI announced a major update, GPT-4o (omni), for their ChatGPT free tier at the most recent announcement event. GPT-4o, as its name implies, is a better version of GPT-4 that has multimodal capabilities.
Chief Technology Officer Mira Murati of OpenAI claims that this new model is far faster and improves audio, visual, and textual skills. She also says that it is more accurate. During a live broadcast that took place on Monday, Murati gave these official statements. The improved model will be available to all users at no extra cost; premium users will have access to five times the capacity limits available to free users.
At the moment, the new GPT-4o is only capable of doing word and image tasks; but, according to a blog post published by OpenAI, all of its capabilities will be “rolled out iteratively.”
Sam Altman, the CEO of OpenAI, said that GPT-4o is totally multimodal, in contrast to GPT 3.5. This implies that it is able to generate content and receive input in the form of text, speech, and pictures. Additionally, Altman said on X that developers who want to test GPT-4o would have access to an API. It is half as costly as GPT-4 Turbo and twice as speedy.
ChatGPT will be able to act as a voice assistant similar to the one shown in the film “Her-like,” which was released in 2013 and is about a conversational AI assistant. This will be possible if ChatGPT acquires further capability for its voice mode. Your camera will allow it to recognize its surroundings and reply to users in real time. Speech mode can only accept voice input and react to one command at a time.