Qwen2.5 14B Instruct logo

Qwen2.5 14B Instruct

Qwen

Qwen2.5-14B-Instruct is an instruction-tuned language model with 14.7 billion parameters and part of the Qwen2.5 series. This model boasts substantial enhancements in instruction adherence, producing extended text (8,000+ tokens), interpreting structured data, and generating JSON outputs. With a context length of 128,000 tokens, it proficiently handles multiple languages, supporting over 29, including Chinese, English, French, and Spanish.

Model Specifications

Technical details and capabilities of Qwen2.5 14B Instruct

Core Specifications

14.7B Parameters

Model size and complexity

18000.0B Training Tokens

Amount of data used in training

131.1K / 8.2K

Input / Output tokens

September 18, 2024

Release date

Capabilities & License

Multimodal Support
Not Supported
Web Hydrated
No
License
apache-2.0

Resources

Research Paper
https://arxiv.org/abs/2407.10671
API Reference
https://www.alibabacloud.com/help/en/model-studio/developer-reference/use-qwen-by-calling-api
Code Repository
https://github.com/QwenLM/Qwen2.5

Performance Insights

Check out how Qwen2.5 14B Instruct handles various AI tasks through comprehensive benchmark results.

100
75
50
25
0
94.8
GSM8K
94.8
(95%)
83.5
HumanEval
83.5
(84%)
82
MBPP
82
(82%)
80
MMLU-Redux
80
(80%)
80
MATH
80
(80%)
79.7
MMLU
79.7
(80%)
78.2
BBH
78.2
(78%)
76.4
MMLU-STEM
76.4
(76%)
72.8
MultiPL-E
72.8
(73%)
67.3
ARC-C
67.3
(67%)
63.7
MMLU-Pro
63.7
(64%)
63.2
MBPP+
63.2
(63%)
58.4
TruthfulQA
58.4
(58%)
51.2
HumanEval+
51.2
(51%)
45.5
GPQA
45.5
(46%)
43
TheoremQA
43
(43%)
GSM8K
HumanEval
MBPP
MMLU-Redux
MATH
MMLU
BBH
MMLU-STEM
MultiPL-E
ARC-C
MMLU-Pro
MBPP+
TruthfulQA
HumanEval+
GPQA
TheoremQA

Detailed Benchmarks

Dive deeper into Qwen2.5 14B Instruct's performance across specific task categories. Expand each section to see detailed metrics and comparisons.

Math

GSM8K

Current model
Other models
Avg (91.9%)

Coding

HumanEval

Current model
Other models
Avg (79.5%)

HumanEval+

Current model
Other models
Avg (62.1%)

MBPP+

Current model
Other models
Avg (65.2%)

Non categorized

BBH

Current model
Other models
Avg (81.1%)

ARC-C

Current model
Other models
Avg (82.3%)

TheoremQA

Current model
Other models
Avg (41.1%)

MMLU-STEM

Current model
Other models
Avg (78.6%)

MultiPL-E

Current model
Other models
Avg (70.3%)

Providers Pricing Coming Soon

We're working on gathering comprehensive pricing data from all major providers for Qwen2.5 14B Instruct. Compare costs across platforms to find the best pricing for your use case.

OpenAI
Anthropic
Google
Mistral AI
Cohere

Share your feedback

Hi, I'm Charlie Palars, the founder of Deepranking.ai. I'm always looking for ways to improve the site and make it more useful for you. You can write me through this form or directly through X at @palarsio.

Your feedback helps us improve our service

Stay Ahead with AI Updates

Get insights on Gemini Pro 2.5, Sonnet 3.7 and more top AI models