Jump to content

GPT-4o: Difference between revisions

From Wikipedia, the free encyclopedia
Content deleted Content added
Line 24: Line 24:


== Capabilities ==
== Capabilities ==
GPT-4o achieves state-of-the-art{{cn|date=May 2024}} results in voice, multilingual, and vision benchmarks, setting new records in audio speech recognition and translation.<ref name="Hello GPT-4o">{{Cite web |title=Hello GPT-4o |url=https://openai.com/index/hello-gpt-4o/ |website=OpenAI}}</ref> GPT-4o scores 88.7 on the Massive Multitask Language Understanding ([[MMLU]]) benchmark compared to 86.5 by GPT-4.<ref name="Hello GPT-4o" /> For voice-to-voice{{emdash}}unlike GPT-3.5 and GPT-4 which convert the voice to text, give the text to the model, then convert the text back to voice using another model{{emdash}}GPT-4o natively supports voice-to-voice making the response near instant and seamless.<ref name="Hello GPT-4o" /><ref>{{Cite web |last=Altman |first=Sam |title= |url=https://twitter.com/sama/status/1790817315069771959 |access-date=16 May 2024 |website=Twitter/X}}</ref>
GPT-4o achieves state-of-the-art<ref>{{cite web |last1=van Rijmenam |first1=Mark |title=OpenAI Launched GPT-4o: The Future of AI Interactions Is Here |url=https://www.thedigitalspeaker.com/openai-gpt4o-future-ai-interactions/ |website=The Digital Speaker |access-date=17 May 2024 |date=13 May 2024}}</ref> results in voice, multilingual, and vision benchmarks, setting new records in audio speech recognition and translation.<ref name="Hello GPT-4o">{{Cite web |title=Hello GPT-4o |url=https://openai.com/index/hello-gpt-4o/ |website=OpenAI}}</ref> GPT-4o scores 88.7 on the Massive Multitask Language Understanding ([[MMLU]]) benchmark compared to 86.5 by GPT-4.<ref name="Hello GPT-4o" /> For voice-to-voice{{emdash}}unlike GPT-3.5 and GPT-4 which convert the voice to text, give the text to the model, then convert the text back to voice using another model{{emdash}}GPT-4o natively supports voice-to-voice making the response near instant and seamless.<ref name="Hello GPT-4o" /><ref>{{Cite web |last=Altman |first=Sam |title= |url=https://twitter.com/sama/status/1790817315069771959 |access-date=16 May 2024 |website=Twitter/X}}</ref>


The model supports over 50 languages,<ref name="TechCrunch" /> covering over 97% of speakers.{{cn|date=May 2024}} Mira Murati demonstrated the model's multilingual capability by speaking Italian to the model and having it translate between English and Italian during the live-streamed OpenAI demo event on 13{{nbsp}}May 2024. In addition, the new tokenizer uses fewer tokens for certain languages, especially languages that are not based on the [[Latin alphabet]], making it cheaper for those languages.<ref name="Hello GPT-4o" />
The model supports over 50 languages,<ref name="TechCrunch" /> covering over 97% of speakers.{{cn|date=May 2024}} Mira Murati demonstrated the model's multilingual capability by speaking Italian to the model and having it translate between English and Italian during the live-streamed OpenAI demo event on 13{{nbsp}}May 2024. In addition, the new tokenizer uses fewer tokens for certain languages, especially languages that are not based on the [[Latin alphabet]], making it cheaper for those languages.<ref name="Hello GPT-4o" />

Revision as of 01:34, 17 May 2024

Generative Pre-trained Transformer 4 Omni (GPT-4o)
Developer(s)OpenAI
Initial releaseMay 13, 2024; 4 months ago (2024-05-13)
PredecessorGPT-4 Turbo
Type
LicenseProprietary
Websiteopenai.com/index/hello-gpt-4o

GPT-4o (GPT-4 omni) is a multilingual, multimodal generative pre-trained transformer designed by OpenAI. It was announced by OpenAI's CTO Mira Murati during a live-streamed demo on 13 May 2024 and released the same day.[1] GPT-4o is free, but with a usage limit that is 5 times higher for ChatGPT Plus subscribers.[2] Its API is twice as fast and half the price of its predecessor, GPT-4 Turbo.[1]

Background

GPT-4o was originally shadow launched on LMSYS, as 3 different models. These 3 models were called gpt2-chatbot, im-a-good-gpt2-chatbot, and im-also-a-good-gpt2-chatbot. On 7 May 2024, Sam Altman revealed that OpenAI was responsible for these mysterious new models.[3]

Capabilities

GPT-4o achieves state-of-the-art[4] results in voice, multilingual, and vision benchmarks, setting new records in audio speech recognition and translation.[5] GPT-4o scores 88.7 on the Massive Multitask Language Understanding (MMLU) benchmark compared to 86.5 by GPT-4.[5] For voice-to-voice—unlike GPT-3.5 and GPT-4 which convert the voice to text, give the text to the model, then convert the text back to voice using another model—GPT-4o natively supports voice-to-voice making the response near instant and seamless.[5][6]

The model supports over 50 languages,[1] covering over 97% of speakers.[citation needed] Mira Murati demonstrated the model's multilingual capability by speaking Italian to the model and having it translate between English and Italian during the live-streamed OpenAI demo event on 13 May 2024. In addition, the new tokenizer uses fewer tokens for certain languages, especially languages that are not based on the Latin alphabet, making it cheaper for those languages.[5]

It is currently the leading model in the Large Model Systems Organization (LMSYS) Elo Arena Benchmarks by the University of California, Berkeley.[7]

See also

References

  1. ^ a b c Wiggers, Kyle (2024-05-13). "OpenAI debuts GPT-4o 'omni' model now powering ChatGPT". TechCrunch. Retrieved 2024-05-13.
  2. ^ Field, Hayden (2024-05-13). "OpenAI launches new AI model GPT-4o and desktop version of ChatGPT". CNBC. Retrieved 2024-05-14.
  3. ^ Sam Altman "https://twitter.com/sama/status/1787222050589028528" Twitter, X. Retrieved 14 May 2024.
  4. ^ van Rijmenam, Mark (13 May 2024). "OpenAI Launched GPT-4o: The Future of AI Interactions Is Here". The Digital Speaker. Retrieved 17 May 2024.
  5. ^ a b c d "Hello GPT-4o". OpenAI.
  6. ^ Altman, Sam. Twitter/X https://twitter.com/sama/status/1790817315069771959. Retrieved 16 May 2024. {{cite web}}: Missing or empty |title= (help)
  7. ^ Fedus, William. "GPT-4o is our new state-of-the-art frontier model".