Qwen2.5-Coder 7B Instruct logo

Qwen2.5-Coder 7B Instruct

Qwen

Qwen2.5-Coder is a powerhouse coding model, expertly trained on 5.5 trillion code tokens and fluent in 92 programming languages. Equipped with a 128K context window, it's engineered for top-tier code generation, intelligent completion, and precise repair. Beyond coding, it retains impressive skills in mathematics and general problem-solving. This model truly shines in complex, multi-language projects and demonstrates advanced code reasoning abilities.

Model Specifications

Technical details and capabilities of Qwen2.5-Coder 7B Instruct

Core Specifications

7.0B Parameters

Model size and complexity

5500.0B Training Tokens

Amount of data used in training

128.0K / 128.0K

Input / Output tokens

February 29, 2024

Knowledge cutoff date

September 18, 2024

Release date

Capabilities & License

Multimodal Support
Not Supported
Web Hydrated
No
License
Apache 2.0

Resources

Research Paper
https://arxiv.org/abs/2409.12186
API Reference
https://www.alibabacloud.com/help/en/model-studio/developer-reference/use-qwen-by-calling-api
Code Repository
https://github.com/QwenLM/Qwen2

Performance Insights

Check out how Qwen2.5-Coder 7B Instruct handles various AI tasks through comprehensive benchmark results.

90
68
45
23
0
88.4
HumanEval
88.4
(98%)
83.9
GSM8K
83.9
(93%)
83.5
MBPP
83.5
(93%)
76.8
HellaSwag
76.8
(85%)
72.9
WinoGrande
72.9
(81%)
68
MMLU-Base
68
(76%)
67.6
MMLU
67.6
(75%)
66.6
MMLU-Redux
66.6
(74%)
60.9
ARC-Challenge
60.9
(68%)
56.5
CRUXEval-Input-CoT
56.5
(63%)
56.0
CRUXEval-Output-CoT
56.0
(62%)
55.6
Aider
55.6
(62%)
50.6
TruthfulQA
50.6
(56%)
46.6
MATH
46.6
(52%)
41
BigCodeBench
41
(46%)
40.1
MMLU-Pro
40.1
(45%)
34
STEM
34
(38%)
34
TheoremQA
34
(38%)
18.2
LiveCodeBench
18.2
(20%)
HumanEval
GSM8K
MBPP
HellaSwag
WinoGrande
MMLU-Base
MMLU
MMLU-Redux
ARC-Challenge
CRUXEval-Input-CoT
CRUXEval-Output-CoT
Aider
TruthfulQA
MATH
BigCodeBench
MMLU-Pro
STEM
TheoremQA
LiveCodeBench

Detailed Benchmarks

Dive deeper into Qwen2.5-Coder 7B Instruct's performance across specific task categories. Expand each section to see detailed metrics and comparisons.

Non categorized

TheoremQA

Current model
Other models
Avg (41.1%)

ARC-Challenge

Current model
Other models
Avg (65.7%)

WinoGrande

Current model
Other models
Avg (77.4%)

Aider

Current model
Other models
Avg (63.9%)

Providers Pricing Coming Soon

We're working on gathering comprehensive pricing data from all major providers for Qwen2.5-Coder 7B Instruct. Compare costs across platforms to find the best pricing for your use case.

OpenAI
Anthropic
Google
Mistral AI
Cohere

Share your feedback

Hi, I'm Charlie Palars, the founder of Deepranking.ai. I'm always looking for ways to improve the site and make it more useful for you. You can write me through this form or directly through X at @palarsio.

Your feedback helps us improve our service

Stay Ahead with AI Updates

Get insights on Gemini Pro 2.5, Sonnet 3.7 and more top AI models