Qwen2.5 72B Instruct logo

Qwen2.5 72B Instruct

Qwen

Qwen2.5-72B-Instruct is a 72 billion parameter language model from the Qwen2.5 series, fine-tuned for instruction following. This model excels at producing extended text outputs exceeding 8,000 tokens, demonstrates strong comprehension of structured information like tables, and is adept at generating structured outputs, with a particular focus on JSON format. Furthermore, it boasts multilingual support for more than 29 languages.

Model Specifications

Technical details and capabilities of Qwen2.5 72B Instruct

Core Specifications

72.7B Parameters

Model size and complexity

18000.0B Training Tokens

Amount of data used in training

131.1K / 8.2K

Input / Output tokens

September 18, 2024

Release date

Capabilities & License

Multimodal Support
Not Supported
Web Hydrated
No
License
qwen

Resources

API Reference
https://www.alibabacloud.com/help/en/model-studio/developer-reference/use-qwen-by-calling-api
Code Repository
https://github.com/QwenLM/Qwen2.5

Performance Insights

Check out how Qwen2.5 72B Instruct handles various AI tasks through comprehensive benchmark results.

100
75
50
25
0
95.8
GSM8K
95.8
(96%)
93.5
MT-bench
93.5
(94%)
88.2
MBPP
88.2
(88%)
86.8
MMLU-Redux
86.8
(87%)
86.6
HumanEval
86.6
(87%)
84.1
IFEval
84.1
(84%)
83.1
MATH
83.1
(83%)
81.6
AlignBench
81.6
(82%)
81.2
Arena-Hard
81.2
(81%)
75.1
MultiPL-E
75.1
(75%)
71.1
MMLU-Pro
71.1
(71%)
55.5
LiveCodeBench
55.5
(56%)
52.3
LiveBench
52.3
(52%)
49
GPQA
49
(49%)
GSM8K
MT-bench
MBPP
MMLU-Redux
HumanEval
IFEval
MATH
AlignBench
Arena-Hard
MultiPL-E
MMLU-Pro
LiveCodeBench
LiveBench
GPQA

Model Comparison

See how Qwen2.5 72B Instruct stacks up against other leading models across key performance metrics.

100
80
60
40
20
0
71.1
MMLU-Pro - Qwen2.5 72B Instruct
71.1
(71%)
69
MMLU-Pro - Qwen2.5 32B Instruct
69
(69%)
63.7
MMLU-Pro - Qwen2.5 14B Instruct
63.7
(64%)
73.3
MMLU-Pro - Llama 3.1 405B Instruct
73.3
(73%)
75.8
MMLU-Pro - Gemini 1.5 Pro
75.8
(76%)
67.3
MMLU-Pro - Gemini 1.5 Flash
67.3
(67%)
49
GPQA - Qwen2.5 72B Instruct
49
(49%)
49.5
GPQA - Qwen2.5 32B Instruct
49.5
(50%)
45.5
GPQA - Qwen2.5 14B Instruct
45.5
(46%)
50.7
GPQA - Llama 3.1 405B Instruct
50.7
(51%)
59.1
GPQA - Gemini 1.5 Pro
59.1
(59%)
51
GPQA - Gemini 1.5 Flash
51
(51%)
83.1
MATH - Qwen2.5 72B Instruct
83.1
(83%)
83.1
MATH - Qwen2.5 32B Instruct
83.1
(83%)
80
MATH - Qwen2.5 14B Instruct
80
(80%)
73.8
MATH - Llama 3.1 405B Instruct
73.8
(74%)
86.5
MATH - Gemini 1.5 Pro
86.5
(87%)
77.9
MATH - Gemini 1.5 Flash
77.9
(78%)
95.8
GSM8K - Qwen2.5 72B Instruct
95.8
(96%)
95.9
GSM8K - Qwen2.5 32B Instruct
95.9
(96%)
94.8
GSM8K - Qwen2.5 14B Instruct
94.8
(95%)
96.8
GSM8K - Llama 3.1 405B Instruct
96.8
(97%)
90.8
GSM8K - Gemini 1.5 Pro
90.8
(91%)
86.2
GSM8K - Gemini 1.5 Flash
86.2
(86%)
86.6
HumanEval - Qwen2.5 72B Instruct
86.6
(87%)
88.4
HumanEval - Qwen2.5 32B Instruct
88.4
(88%)
83.5
HumanEval - Qwen2.5 14B Instruct
83.5
(84%)
89
HumanEval - Llama 3.1 405B Instruct
89
(89%)
84.1
HumanEval - Gemini 1.5 Pro
84.1
(84%)
74.3
HumanEval - Gemini 1.5 Flash
74.3
(74%)
MMLU-Pro
GPQA
MATH
GSM8K
HumanEval
Qwen2.5 72B Instruct
Qwen2.5 32B Instruct
Qwen2.5 14B Instruct
Llama 3.1 405B Instruct
Gemini 1.5 Pro
Gemini 1.5 Flash

Detailed Benchmarks

Dive deeper into Qwen2.5 72B Instruct's performance across specific task categories. Expand each section to see detailed metrics and comparisons.

Math

GSM8K

Current model
Other models
Avg (93.1%)

Knowledge

Non categorized

MMLU-Pro

Current model
Other models
Avg (69.6%)

MultiPL-E

Current model
Other models
Avg (70.3%)

LiveBench

Current model
Other models
Avg (54.7%)

IFEval

Current model
Other models
Avg (83.6%)

Arena-Hard

Current model
Other models
Avg (66.6%)

AlignBench

Current model
Other models
Avg (76.9%)

MT-bench

Current model
Other models
Avg (90.5%)

Providers Pricing Coming Soon

We're working on gathering comprehensive pricing data from all major providers for Qwen2.5 72B Instruct. Compare costs across platforms to find the best pricing for your use case.

OpenAI
Anthropic
Google
Mistral AI
Cohere

Share your feedback

Hi, I'm Charlie Palars, the founder of Deepranking.ai. I'm always looking for ways to improve the site and make it more useful for you. You can write me through this form or directly through X at @palarsio.

Your feedback helps us improve our service

Stay Ahead with AI Updates

Get insights on Gemini Pro 2.5, Sonnet 3.7 and more top AI models