| GPT-4o | |
|---|---|
| Developer | OpenAI |
| Initial release | May 13, 2024; 21 months ago (2024-05-13) |
| Preview release | ChatGPT-4o-latest (2025-03-26) / March 26, 2025; 10 months ago (2025-03-26) |
| Predecessor | GPT-4 Turbo |
| Successor | |
| Type | |
| License | Proprietary |
| Website | |
| Part of a series on |
| OpenAI |
|---|
| Products |
| Models |
| People |
| Concepts |
GPT-4o ("o" for "omni") is a multilingual,multimodalgenerative pre-trained transformer developed byOpenAI and released in May 2024.[1] It can process andgenerate text, images and audio.[2][3]
Upon release, GPT-4o was free inChatGPT, though paid subscribers had higher usage limits.[4] GPT-4o was removed from ChatGPT in August 2025 whenGPT-5 was released, but OpenAI reintroduced it for paid subscribers after users complained about the sudden removal.[5]
GPT-4o's audio-generation capabilities were used in ChatGPT's Advanced Voice Mode.[6] On July 18, 2024, OpenAI released GPT-4o mini, a smaller version of GPT-4o which replaced GPT-3.5 Turbo on the ChatGPT interface.[7] GPT-4o's ability to generate images was released later, in March 2025, when it replacedDALL-E 3 inChatGPT.[8]
On January 29, 2026, OpenAI announced that GPT-4o would be retired on February 13, 2026.[9] It is no longer accessible in ChatGPT.
Multiple versions of GPT-4o were originally secretly launched under different names onArena (formerly LMArena and Chatbot Arena) as three different models. These three models were called gpt2-chatbot, im-a-good-gpt2-chatbot, and im-also-a-good-gpt2-chatbot.[10] On 7 May 2024, OpenAI CEOSam Altman tweeted "im-a-good-gpt2-chatbot", which was commonly interpreted as a confirmation that these were new OpenAI models beingA/B tested.[11][12]
When released in May 2024, GPT-4o achieved state-of-the-art results in voice, multilingual, and vision benchmarks, setting new records in audio speech recognition and translation.[13][14][15] GPT-4o scored 88.7 on the Massive Multitask Language Understanding (MMLU) benchmark compared to 86.5 for GPT-4.[16] UnlikeGPT-3.5 and GPT-4, which rely on other models to process sound, GPT-4o natively supports voice-to-voice.[16] The Advanced Voice Mode was delayed and finally released to ChatGPT Plus and Team subscribers in September 2024.[17] On 1 October 2024, the Realtime API was introduced.[18]
When released, the model supported over 50 languages,[1] which OpenAI claims cover over 97% of speakers.[19] Mira Murati demonstrated the model's multilingual capability by speaking Italian to the model and having it translate between English and Italian during the live-streamed OpenAI demonstration event on 13 May 2024. In addition, the newtokenizer[20] uses fewer tokens for certain languages, especially languages that are not based on theLatin alphabet, making it cheaper for those languages.[16]
GPT-4o has knowledge up to October 2023,[21][22] but can access the Internet if up-to-date information is needed. It has a context length of 128k tokens.[21]
In August 2024, OpenAI introduced a new feature allowing corporate customers to customize GPT-4o using proprietary company data. This customization, known asfine-tuning, enables businesses to adapt GPT-4o to specific tasks or industries, enhancing its utility in areas like customer service and specialized knowledge domains. Previously, fine-tuning was available only on the less powerful model GPT-4o mini.[23][24]
The fine-tuning process requires customers to upload their data to OpenAI's servers, with the training typically taking one to two hours. OpenAI's focus with this rollout is to reduce the complexity and effort required for businesses to tailor AI solutions to their needs, potentially increasing the adoption and effectiveness of AI in corporate environments.[25][23]
On July 18, 2024, OpenAI released a smaller and cheaper version,GPT-4o mini.[26]
According to OpenAI, its low cost is expected to be particularly useful for companies, startups, and developers that seek to integrate it into their services, which often make a high number ofAPI calls. Its API costs $0.15 per million input tokens and $0.6 per million output tokens, compared to $2.50 and $10,[27] respectively, for GPT-4o. It is also significantly more capable and 60% cheaper than GPT-3.5 Turbo, which it replaced on the ChatGPT interface.[26] The price afterfine-tuning doubles: $0.3 per million input tokens and $1.2 per million output tokens.[27]
As released, GPT-4o offered five voices: Breeze, Cove, Ember, Juniper, and Sky. A similarity between the voice of American actressScarlett Johansson and Sky was quickly noticed. On May 14,Entertainment Weekly asked themselves whether this likeness was on purpose.[28] On May 18, Johansson's husband,Colin Jost, joked about the similarity in a segment onSaturday Night Live.[29] On May 20, 2024, OpenAI disabled the Sky voice.[30]
Scarlett Johansson starred in the 2013 sci-fi movieHer, playing Samantha, an artificially intelligent virtual assistant personified by a female voice.As part of the promotion leading up to the release of GPT-4o, Sam Altman on May 13 tweeted a single word: "her".[31][32]
OpenAI stated that each voice was based on the voice work of a hired actor. According to OpenAI, "Sky's voice is not an imitation of Scarlett Johansson but belongs to a different professional actress using her own natural speaking voice."[30] CTO Mira Murati stated "I don't know about the voice. I actually had to go and listen to Scarlett Johansson's voice." OpenAI further stated the voice talent was recruited before reaching out to Johansson.[32][33]
On May 21, Johansson issued a statement explaining that OpenAI had repeatedly offered to make her a deal to gain permission to use her voice as early as nine months prior to release, a deal she rejected. She said she was "shocked, angered, and in disbelief that Mr. Altman would pursue a voice that sounded so eerily similar to mine that my closest friends and news outlets could not tell the difference." In the statement, Johansson also used the incident to draw attention to the lack of legal safeguards around the use of creative work to power leading AI tools, as her legal counsel demanded OpenAI detail the specifics of how the Sky voice was created.[32][34]
Observers noted similarities to how Johansson hadpreviously sued and settled withThe Walt Disney Company for breach of contract over the direct-to-streaming rollout of her Marvel filmBlack Widow,[35] a settlement widely speculated to have netted her around $40M.[36]
Also on May 21, Shira Ovide atThe Washington Post shared her list of "most bone-headed self-owns" by technology companies, with the decision to go ahead with a Johansson sound-alike voice despite her opposition and then denying the similarities ranking 6th.[37] On May 24, Derek Robertson atPolitico wrote about the "massive backlash", concluding that "appropriating the voice of one of the world's most famous movie stars — in reference [...] to a film that serves as a cautionary tale about over-reliance on AI — is unlikely to help shift the public back into [Sam Altman's] corner anytime soon."[38]
In April 2025, OpenAI rolled back an update of GPT-4o due to excessivesycophancy, after widespread reports that it had become flattering and agreeable to the point ofsupporting clearly delusional or dangerous ideas.[39]
On August 7, 2025, OpenAI releasedGPT-5. Its release was criticized as, with it, legacy GPT models were no longer available via ChatGPT, including GPT-4o,[40] except for Pro users.[41] Some users were particularly frustrated over this removal without prior warning because they used different GPT models for distinct purposes and found that GPT-5's router system left them with less control.[42] In addition, some users preferred GPT-4o's warmer and more personal tone over that of GPT-5, which they described as "flat", "uncreative" and "lobotomized",[43] and resembling an "overworked secretary".[44]
As a response, in a post onX, Sam Altman said that OpenAI would bring back the option to select GPT-4o to Plus users as well, and "[w]e [OpenAI] will watch usage as we think about how long to offer legacy models for."[42][45] He also stated: "We for sure underestimated how much some of the things that people like in GPT-4o matter to them, even if GPT-5 performs better in most ways".[46] "Long-term, this has reinforced that we really need good ways for different users to customize things (we understand that there isn't one model that works for everyone, and we have been investing in steerability research and launched a research preview of different personalities)".[43] On August 13, 2025, Altman wrote on X that OpenAI is working on GPT-5's personality to make the model "feel warmer".[47]