Claude 3 Sonnet logo

Claude 3 Sonnet

Anthropic

Claude 3 Sonnet is designed to provide the optimal combination of intelligence and speed, making it perfect for handling demanding business tasks. It offers exceptional performance at a more economical price point than similar models and is built to reliably handle extensive AI deployments.

Model Specifications

Technical details and capabilities of Claude 3 Sonnet

Core Specifications

200.0K / 200.0K

Input / Output tokens

February 28, 2024

Release date

Capabilities & License

Multimodal Support
Supported
Web Hydrated
No
License
Proprietary

Resources

Research Paper
https://www-cdn.anthropic.com/de8ba9b01c9ab7cbabf5c33b80b7bbc618857627/Model_Card_Claude_3.pdf
API Reference
https://www.anthropic.com/claude
Playground
https://claude.ai

Performance Insights

Check out how Claude 3 Sonnet handles various AI tasks through comprehensive benchmark results.

100
75
50
25
0
93.2
ARC Challenge
93.2
(93%)
92.3
GSM8K
92.3
(92%)
89
HellaSwag
89
(89%)
83.5
MGSM
83.5
(84%)
82.9
BIG-Bench-Hard
82.9
(83%)
79
MMLU
79
(79%)
78.9
DROP
78.9
(79%)
73
HumanEval
73
(73%)
56.8
MMLU-Pro
56.8
(57%)
43.1
MATH
43.1
(43%)
40.4
GPQA
40.4
(40%)
ARC Challenge
GSM8K
HellaSwag
MGSM
BIG-Bench-Hard
MMLU
DROP
HumanEval
MMLU-Pro
MATH
GPQA

Model Comparison

See how Claude 3 Sonnet stacks up against other leading models across key performance metrics.

90
72
54
36
18
0
79
MMLU - Claude 3 Sonnet
79
(88%)
81.3
MMLU - Grok-1.5
81.3
(90%)
78.9
MMLU - Phi-3.5-MoE-instruct
78.9
(88%)
69
MMLU - Phi-3.5-mini-instruct
69
(77%)
82.3
MMLU - Qwen2 72B Instruct
82.3
(91%)
86.8
MMLU - Claude 3 Opus
86.8
(96%)
40.4
GPQA - Claude 3 Sonnet
40.4
(45%)
35.9
GPQA - Grok-1.5
35.9
(40%)
36.8
GPQA - Phi-3.5-MoE-instruct
36.8
(41%)
30.4
GPQA - Phi-3.5-mini-instruct
30.4
(34%)
42.4
GPQA - Qwen2 72B Instruct
42.4
(47%)
50.4
GPQA - Claude 3 Opus
50.4
(56%)
43.1
MATH - Claude 3 Sonnet
43.1
(48%)
50.6
MATH - Grok-1.5
50.6
(56%)
59.5
MATH - Phi-3.5-MoE-instruct
59.5
(66%)
48.5
MATH - Phi-3.5-mini-instruct
48.5
(54%)
59.7
MATH - Qwen2 72B Instruct
59.7
(66%)
60.1
MATH - Claude 3 Opus
60.1
(67%)
73
HumanEval - Claude 3 Sonnet
73
(81%)
74.1
HumanEval - Grok-1.5
74.1
(82%)
70.7
HumanEval - Phi-3.5-MoE-instruct
70.7
(79%)
62.8
HumanEval - Phi-3.5-mini-instruct
62.8
(70%)
86
HumanEval - Qwen2 72B Instruct
86
(96%)
84.9
HumanEval - Claude 3 Opus
84.9
(94%)
56.8
MMLU-Pro - Claude 3 Sonnet
56.8
(63%)
51
MMLU-Pro - Grok-1.5
51
(57%)
54.3
MMLU-Pro - Phi-3.5-MoE-instruct
54.3
(60%)
47.4
MMLU-Pro - Phi-3.5-mini-instruct
47.4
(53%)
64.4
MMLU-Pro - Qwen2 72B Instruct
64.4
(72%)
68.5
MMLU-Pro - Claude 3 Opus
68.5
(76%)
MMLU
GPQA
MATH
HumanEval
MMLU-Pro
Claude 3 Sonnet
Grok-1.5
Phi-3.5-MoE-instruct
Phi-3.5-mini-instruct
Qwen2 72B Instruct
Claude 3 Opus

Detailed Benchmarks

Dive deeper into Claude 3 Sonnet's performance across specific task categories. Expand each section to see detailed metrics and comparisons.

Reasoning

DROP

Current model
Other models
Avg (77.8%)

HellaSwag

Current model
Other models
Avg (87.0%)

Knowledge

Non categorized

MGSM

Current model
Other models
Avg (80.9%)

BIG-Bench-Hard

Current model
Other models
Avg (81.1%)

Providers Pricing Coming Soon

We're working on gathering comprehensive pricing data from all major providers for Claude 3 Sonnet. Compare costs across platforms to find the best pricing for your use case.

OpenAI
Anthropic
Google
Mistral AI
Cohere

Share your feedback

Hi, I'm Charlie Palars, the founder of Deepranking.ai. I'm always looking for ways to improve the site and make it more useful for you. You can write me through this form or directly through X at @palarsio.

Your feedback helps us improve our service

Stay Ahead with AI Updates

Get insights on Gemini Pro 2.5, Sonnet 3.7 and more top AI models