GPT-4 Turbo

OpenAI

The newest GPT-4 model delivers enhanced performance, up-to-date knowledge, and expanded capabilities. Benefit from quicker response times and more cost-effective pricing compared to earlier iterations.

Model Specifications

Technical details and capabilities of GPT-4 Turbo

Core Specifications

128.0K / 4.1K

Input / Output tokens

December 30, 2023

Knowledge cutoff date

April 8, 2024

Release date

Capabilities & License

Multimodal Support
Not Supported
Web Hydrated
Yes
License
Proprietary

Resources

API Reference
https://platform.openai.com/docs/models/gpt-4-turbo-and-gpt-4
Playground
https://platform.openai.com/playground

Performance Insights

Check out how GPT-4 Turbo handles various AI tasks through comprehensive benchmark results.

90
68
45
23
0
88.5
MGSM
88.5
(98%)
87.1
HumanEval
87.1
(97%)
86.5
MMLU
86.5
(96%)
86
DROP
86
(96%)
72.6
MATH
72.6
(81%)
48
GPQA
48
(53%)
MGSM
HumanEval
MMLU
DROP
MATH
GPQA

Model Comparison

See how GPT-4 Turbo stacks up against other leading models across key performance metrics.

100
80
60
40
20
0
86.5
MMLU - GPT-4 Turbo
86.5
(87%)
86
MMLU - Llama 3.3 70B Instruct
86
(86%)
88.7
MMLU - GPT-4o
88.7
(89%)
82
MMLU - GPT-4o mini
82
(82%)
86.8
MMLU - Claude 3 Opus
86.8
(87%)
90.4
MMLU - Claude 3.5 Sonnet
90.4
(90%)
48
GPQA - GPT-4 Turbo
48
(48%)
50.5
GPQA - Llama 3.3 70B Instruct
50.5
(51%)
53.6
GPQA - GPT-4o
53.6
(54%)
40.2
GPQA - GPT-4o mini
40.2
(40%)
50.4
GPQA - Claude 3 Opus
50.4
(50%)
59.4
GPQA - Claude 3.5 Sonnet
59.4
(59%)
72.6
MATH - GPT-4 Turbo
72.6
(73%)
77
MATH - Llama 3.3 70B Instruct
77
(77%)
76.6
MATH - GPT-4o
76.6
(77%)
70.2
MATH - GPT-4o mini
70.2
(70%)
60.1
MATH - Claude 3 Opus
60.1
(60%)
71.1
MATH - Claude 3.5 Sonnet
71.1
(71%)
87.1
HumanEval - GPT-4 Turbo
87.1
(87%)
88.4
HumanEval - Llama 3.3 70B Instruct
88.4
(88%)
90.2
HumanEval - GPT-4o
90.2
(90%)
87.2
HumanEval - GPT-4o mini
87.2
(87%)
84.9
HumanEval - Claude 3 Opus
84.9
(85%)
92
HumanEval - Claude 3.5 Sonnet
92
(92%)
88.5
MGSM - GPT-4 Turbo
88.5
(89%)
91.1
MGSM - Llama 3.3 70B Instruct
91.1
(91%)
90.5
MGSM - GPT-4o
90.5
(91%)
87
MGSM - GPT-4o mini
87
(87%)
90.7
MGSM - Claude 3 Opus
90.7
(91%)
91.6
MGSM - Claude 3.5 Sonnet
91.6
(92%)
MMLU
GPQA
MATH
HumanEval
MGSM
GPT-4 Turbo
Llama 3.3 70B Instruct
GPT-4o
GPT-4o mini
Claude 3 Opus
Claude 3.5 Sonnet

Detailed Benchmarks

Dive deeper into GPT-4 Turbo's performance across specific task categories. Expand each section to see detailed metrics and comparisons.

Coding

HumanEval

Current model
Other models
Avg (82.5%)

Reasoning

DROP

Current model
Other models
Avg (84.1%)

Knowledge

MMLU

Current model
Other models
Avg (84.9%)

MATH

Current model
Other models
Avg (71.0%)

Non categorized

MGSM

Current model
Other models
Avg (85.1%)

Providers Pricing Coming Soon

We're working on gathering comprehensive pricing data from all major providers for GPT-4 Turbo. Compare costs across platforms to find the best pricing for your use case.

OpenAI
Anthropic
Google
Mistral AI
Cohere

Share your feedback

Hi, I'm Charlie Palars, the founder of Deepranking.ai. I'm always looking for ways to improve the site and make it more useful for you. You can write me through this form or directly through X at @palarsio.

Your feedback helps us improve our service

Stay Ahead with AI Updates

Get insights on Gemini Pro 2.5, Sonnet 3.7 and more top AI models