GPT-4o: OpenAI Unveils Its Latest Language Model, Available for Free to Users
After a ton of speculation on social media and other forums about what OpenAI has in store for us, yesterday, OpenAI finally revealed their latest and most powerful LLM to date — GPT-4o (‘o’ for omni). In case you missed the launch event of GPT-4o, let’s go over the capabilities of GPT-4o and the features it offers.
Enhanced Audio, Text and Vision Capabilites
GPT-4 Turbo is a powerful model, but it comes with one drawback — latency. When compared to GPT-3.5 Turbo, GPT-4 Turbo is still considerably slower. GPT-4o addresses this drawback and is 2x faster than GPT-4 Turbo. This opens up a broader spectrum of use cases involving the integration of data from speech, text, and vision, taking it one step further from multi-modal to omni-modal. The main difference between multi-modal and omni-modal is that in omni-modal, all three sources can be seamlessly run in parallel.
These enhancements also enable the model to generate speech with improved voice modulation, the capability to understand sarcasm, and enhanced natural conversational abilities.
Reduced pricing and available for free to ChatGPT users
Although GPT-4o is more efficient and faster compared to the outgoing GPT-4 Turbo, it is half the price (API) of GPT-4 Turbo, meaning that GPT-4o will cost US$5.00/1M input tokens and US$15.00/1M output tokens. With the better pricing, the context window is now 128k tokens, and the knowledge cutoff is October 2023.
As a cherry on top, GPT-4o will be available to all ChatGPT users for free (ChatGPT Plus users will have 5x cap for GPT-4o). Alongside this, OpenAI also unveiled the ChatGPT desktop app, which will allow users to make use of the vision capabilities of GPT-4o to read and comprehend the content being displayed on the screen. Users will also be able to talk to ChatGPT using the desktop app.
GPT-4o Demo
OpenAI stated that they are rolling out access to GPT-4o in stages over the next few weeks, with ChatGPT Plus users receiving priority and early access to the model. We will understand the true potential of this model only once we get access to it in the coming weeks. Exciting times ahead!