GPT-4o

OpenAI

GPT-4o ("o" for "omni") is a versatile AI model capable of processing various inputs, including text, audio, images, and video, and producing outputs in text, audio, and image formats. It delivers the same high performance as GPT-4 Turbo in text and code-related tasks, while offering enhanced capabilities in understanding non-English languages, visual content, and audio.

Model Specifications

Technical details and capabilities of GPT-4o

Core Specifications

128.0K / 16.4K

Input / Output tokens

August 5, 2024

Release date

Capabilities & License

Multimodal Support
Supported
Web Hydrated
No
License
Proprietary

Resources

API Reference
https://platform.openai.com/docs/api-reference
Playground
https://chat.openai.com/

Performance Insights

Check out how GPT-4o handles various AI tasks through comprehensive benchmark results.

100
75
50
25
0
94.2
AI2D
94.2
(94%)
92.8
DocVQA
92.8
(93%)
90.5
MGSM
90.5
(91%)
90.2
HumanEval
90.2
(90%)
90
RepoQA 32k
90
(90%)
88.7
MMLU
88.7
(89%)
88
MMLU
88
(88%)
88
MBPPPlus
88
(88%)
85.7
ChartQA
85.7
(86%)
84
IFEval
84
(84%)
83.4
DROP
83.4
(83%)
76.6
MATH
76.6
(77%)
74.7
MMLU-Pro
74.7
(75%)
74
BFCL
74
(74%)
72.6
MMLU-Pro
72.6
(73%)
72.2
EgoSchema
72.2
(72%)
70
SQL
70
(70%)
69.1
MMMU
69.1
(69%)
63.8
MathVista
63.8
(64%)
63.8
MathVista
63.8
(64%)
61.9
ActivityNet
61.9
(62%)
61.8
SimpleQA
61.8
(62%)
60
Taubench Retail
60
(60%)
53.6
GPQA
53.6
(54%)
53.6
GPQA
53.6
(54%)
41
Taubench Airline
41
(41%)
27.1
Aider Polyglot
27.1
(27%)
13.4
AIME 2024
13.4
(13%)
11
Codeforces
11
(11%)
AI2D
DocVQA
MGSM
HumanEval
RepoQA 32k
MMLU
MMLU
MBPPPlus
ChartQA
IFEval
DROP
MATH
MMLU-Pro
BFCL
MMLU-Pro
EgoSchema
SQL
MMMU
MathVista
MathVista
ActivityNet
SimpleQA
Taubench Retail
GPQA
GPQA
Taubench Airline
Aider Polyglot
AIME 2024
Codeforces

Detailed Benchmarks

Dive deeper into GPT-4o's performance across specific task categories. Expand each section to see detailed metrics and comparisons.

Math

AIME 2024

Current model
Other models
Avg (62.3%)

Coding

Codeforces

90.0%
79.0%
68.0%
47.0%
41.3%
11.0%
Current model
Other models
Avg (52.4%)

Aider Polyglot

Current model
Other models
Avg (51.2%)

Reasoning

DROP

Current model
Other models
Avg (82.5%)

Knowledge

MMLU

Current model
Other models
Avg (87.0%)

GPQA

Current model
Other models
Avg (55.3%)

MATH

Current model
Other models
Avg (74.3%)

Non categorized

MGSM

Current model
Other models
Avg (86.0%)

MathVista

Current model
Other models
Avg (58.6%)

MMLU-Pro

Current model
Other models
Avg (72.0%)

MMMU

Current model
Other models
Avg (63.6%)

AI2D

Current model
Other models
Avg (90.5%)

DocVQA

Current model
Other models
Avg (91.7%)

EgoSchema

Current model
Other models
Avg (66.5%)

SimpleQA

Current model
Other models
Avg (45.1%)

IFEval

Current model
Other models
Avg (83.2%)

Taubench Retail

Current model
Other models
Avg (58.3%)

Taubench Airline

Current model
Other models
Avg (38.0%)

BFCL

Current model
Other models
Avg (71.8%)

MBPPPlus

Current model
Other models
Avg (89.3%)

SQL

Current model
Other models
Avg (66.3%)

RepoQA 32k

Current model
Other models
Avg (89.0%)

Providers Pricing Coming Soon

We're working on gathering comprehensive pricing data from all major providers for GPT-4o. Compare costs across platforms to find the best pricing for your use case.

OpenAI
Anthropic
Google
Mistral AI
Cohere

Share your feedback

Hi, I'm Charlie Palars, the founder of Deepranking.ai. I'm always looking for ways to improve the site and make it more useful for you. You can write me through this form or directly through X at @palarsio.

Your feedback helps us improve our service

Stay Ahead with AI Updates

Get insights on Gemini Pro 2.5, Sonnet 3.7 and more top AI models