Qwen2 7B Instruct logo

Qwen2 7B Instruct

Qwen

Qwen2-7B-Instruct is a 7 billion parameter language model finely tuned for instruction following. It boasts an extensive context window, capable of processing sequences up to 131,072 tokens in length.

Model Specifications

Technical details and capabilities of Qwen2 7B Instruct

Core Specifications

7.6B Parameters

Model size and complexity

131.1K / 131.1K

Input / Output tokens

July 22, 2024

Release date

Capabilities & License

Multimodal Support
Not Supported
Web Hydrated
No
License
apache-2.0

Resources

Research Paper
https://arxiv.org/abs/2309.00071
API Reference
https://huggingface.co/Qwen/Qwen2-7B-Instruct
Playground
https://huggingface.co/Qwen/Qwen2-7B-Instruct
Code Repository
https://huggingface.co/Qwen/Qwen2-7B-Instruct

Performance Insights

Check out how Qwen2 7B Instruct handles various AI tasks through comprehensive benchmark results.

90
68
45
23
0
84.1
MT-Bench
84.1
(93%)
82.3
GSM8K
82.3
(91%)
79.9
Humaneval
79.9
(89%)
77.2
C-Eval
77.2
(86%)
72.1
AlignBench
72.1
(80%)
70.5
MMLU
70.5
(78%)
70.3
Evalplus
70.3
(78%)
67.2
MBPP
67.2
(75%)
59.1
MultiPL-E
59.1
(66%)
49.6
MATH
49.6
(55%)
44.1
MMLU-Pro
44.1
(49%)
26.6
LiveCodeBench
26.6
(30%)
25.3
GPQA
25.3
(28%)
25.3
TheroemQA
25.3
(28%)
MT-Bench
GSM8K
Humaneval
C-Eval
AlignBench
MMLU
Evalplus
MBPP
MultiPL-E
MATH
MMLU-Pro
LiveCodeBench
GPQA
TheroemQA

Model Comparison

See how Qwen2 7B Instruct stacks up against other leading models across key performance metrics.

100
80
60
40
20
0
70.5
MMLU - Qwen2 7B Instruct
70.5
(71%)
69
MMLU - Phi-3.5-mini-instruct
69
(69%)
81.3
MMLU - Grok-1.5
81.3
(81%)
78.9
MMLU - Phi-3.5-MoE-instruct
78.9
(79%)
79
MMLU - Claude 3 Sonnet
79
(79%)
82.3
MMLU - Qwen2 72B Instruct
82.3
(82%)
44.1
MMLU-Pro - Qwen2 7B Instruct
44.1
(44%)
47.4
MMLU-Pro - Phi-3.5-mini-instruct
47.4
(47%)
51
MMLU-Pro - Grok-1.5
51
(51%)
54.3
MMLU-Pro - Phi-3.5-MoE-instruct
54.3
(54%)
56.8
MMLU-Pro - Claude 3 Sonnet
56.8
(57%)
64.4
MMLU-Pro - Qwen2 72B Instruct
64.4
(64%)
25.3
GPQA - Qwen2 7B Instruct
25.3
(25%)
30.4
GPQA - Phi-3.5-mini-instruct
30.4
(30%)
35.9
GPQA - Grok-1.5
35.9
(36%)
36.8
GPQA - Phi-3.5-MoE-instruct
36.8
(37%)
40.4
GPQA - Claude 3 Sonnet
40.4
(40%)
42.4
GPQA - Qwen2 72B Instruct
42.4
(42%)
82.3
GSM8K - Qwen2 7B Instruct
82.3
(82%)
86.2
GSM8K - Phi-3.5-mini-instruct
86.2
(86%)
90
GSM8K - Grok-1.5
90
(90%)
88.7
GSM8K - Phi-3.5-MoE-instruct
88.7
(89%)
92.3
GSM8K - Claude 3 Sonnet
92.3
(92%)
91.1
GSM8K - Qwen2 72B Instruct
91.1
(91%)
49.6
MATH - Qwen2 7B Instruct
49.6
(50%)
48.5
MATH - Phi-3.5-mini-instruct
48.5
(49%)
50.6
MATH - Grok-1.5
50.6
(51%)
59.5
MATH - Phi-3.5-MoE-instruct
59.5
(60%)
43.1
MATH - Claude 3 Sonnet
43.1
(43%)
59.7
MATH - Qwen2 72B Instruct
59.7
(60%)
MMLU
MMLU-Pro
GPQA
GSM8K
MATH
Qwen2 7B Instruct
Phi-3.5-mini-instruct
Grok-1.5
Phi-3.5-MoE-instruct
Claude 3 Sonnet
Qwen2 72B Instruct

Detailed Benchmarks

Dive deeper into Qwen2 7B Instruct's performance across specific task categories. Expand each section to see detailed metrics and comparisons.

Math

Coding

MBPP

Current model
Other models
Avg (71.7%)

LiveCodeBench

Current model
Other models
Avg (32.1%)

Non categorized

TheroemQA

Current model
Other models
Avg (34.9%)

MultiPL-E

Current model
Other models
Avg (70.3%)

C-Eval

Current model
Other models
Avg (84.8%)

AlignBench

Current model
Other models
Avg (76.9%)

Providers Pricing Coming Soon

We're working on gathering comprehensive pricing data from all major providers for Qwen2 7B Instruct. Compare costs across platforms to find the best pricing for your use case.

OpenAI
Anthropic
Google
Mistral AI
Cohere

Share your feedback

Hi, I'm Charlie Palars, the founder of Deepranking.ai. I'm always looking for ways to improve the site and make it more useful for you. You can write me through this form or directly through X at @palarsio.

Your feedback helps us improve our service

Stay Ahead with AI Updates

Get insights on Gemini Pro 2.5, Sonnet 3.7 and more top AI models