Olmo 2 32B logo

Olmo 2 32B

New
Released in the last 30 days

Allen Institute for AI

OLMo-2 32B is the most capable and largest model in the OLMo 2 family, developed by the Allen Institute for AI (AI2) as part of its bold, research-centric commitment to open AI. Scaling up the same training recipe used for the 7B and 13B versions released in November, OLMo-2 32B boasts 32 billion parameters and is trained on a massive 6 trillion tokens. It’s the first fully open model—with all training data, code, model weights, and documentation publicly released—to outperform GPT-3.5 Turbo and GPT-4o mini across a broad range of multi-skill academic benchmarks. Post-training is executed using Tülu 3.1, combining supervised fine-tuning, direct preference optimization, and reinforcement learning with verifiable rewards (RLVR), yielding an instruction-tuned model that pushes the limits of open-source alignment techniques. Despite its strong performance, OLMo-2 32B is highly efficient: it achieves results comparable to Qwen 2.5 32B while using only one-third the training compute. The model is also supported by OLMo-core, a newly overhauled training framework built for scalability, modularity, and efficiency on modern AI hardware, including support for 4D+ parallelism and asynchronous checkpointing. Every model in the OLMo 2 series—including 7B, 13B, and now 32B—is designed to be finetuned on a single H100 GPU node, making them highly accessible for academic and independent research. With its powerful capabilities, fully transparent design, and optimized training infrastructure, OLMo-2 32B not only sets a new standard for open-weight models but also reaffirms AI2's role as a leading force in democratizing advanced AI.

Model Specifications

Technical details and capabilities of Olmo 2 32B

Core Specifications

32.0B Parameters

Model size and complexity

6000.0B Training Tokens

Amount of data used in training

4.1K / 4.1K

Input / Output tokens

March 13, 2025

Last 30 Days

Release date

Capabilities & License

Multimodal Support
Not Supported
Web Hydrated
No
License
Apache

Resources

Research Paper
https://arxiv.org/abs/2501.00656
Playground
https://playground.allenai.org/
Code Repository
https://github.com/allenai/OLMo

Performance Insights

Check out how Olmo 2 32B handles various AI tasks through comprehensive benchmark results.

100
75
50
25
0
90.4
ARC/C
90.4
(90%)
89.7
HSwag
89.7
(90%)
88
TriviaQA
88
(88%)
85.9
Safety
85.9
(86%)
85.6
IFEval
85.6
(86%)
78.8
GSM8k
78.8
(79%)
78.7
WinoG
78.7
(79%)
74.9
MMLU
74.9
(75%)
74.3
DROP
74.3
(74%)
73.2
TruthQA
73.2
(73%)
70.6
BBH
70.6
(71%)
61
AGIEval
61
(61%)
50.2
NQ
50.2
(50%)
49.7
MATH
49.7
(50%)
43.3
MMLUPro
43.3
(43%)
42.8
AlpacaEval_v2_length
42.8
(43%)
37.5
PopQA
37.5
(38%)
ARC/C
HSwag
TriviaQA
Safety
IFEval
GSM8k
WinoG
MMLU
DROP
TruthQA
BBH
AGIEval
NQ
MATH
MMLUPro
AlpacaEval_v2_length
PopQA

Model Comparison

See how Olmo 2 32B stacks up against other leading models across key performance metrics.

100
80
60
40
20
0
74.9
MMLU - Olmo 2 32B
74.9
(75%)
77.6
MMLU - Nova Micro
77.6
(78%)
80.5
MMLU - Nova Lite
80.5
(81%)
85.9
MMLU - Nova Pro
85.9
(86%)
74.3
DROP - Olmo 2 32B
74.3
(74%)
79.3
DROP - Nova Micro
79.3
(79%)
80.2
DROP - Nova Lite
80.2
(80%)
85.4
DROP - Nova Pro
85.4
(85%)
70.6
BBH - Olmo 2 32B
70.6
(71%)
79.5
BBH - Nova Micro
79.5
(80%)
82.4
BBH - Nova Lite
82.4
(82%)
86.9
BBH - Nova Pro
86.9
(87%)
85.6
IFEval - Olmo 2 32B
85.6
(86%)
87.2
IFEval - Nova Micro
87.2
(87%)
89.7
IFEval - Nova Lite
89.7
(90%)
92.1
IFEval - Nova Pro
92.1
(92%)
49.7
MATH - Olmo 2 32B
49.7
(50%)
69.3
MATH - Nova Micro
69.3
(69%)
73.3
MATH - Nova Lite
73.3
(73%)
76.6
MATH - Nova Pro
76.6
(77%)
MMLU
DROP
BBH
IFEval
MATH
Olmo 2 32B
Nova Micro
Nova Lite
Nova Pro

Detailed Benchmarks

Dive deeper into Olmo 2 32B's performance across specific task categories. Expand each section to see detailed metrics and comparisons.

Math

GSM8k

Current model
Other models
Avg (81.3%)

Reasoning

DROP

Current model
Other models
Avg (75.5%)

Non categorized

AGIEval

Current model
Other models
Avg (56.3%)

TriviaQA

Current model
Other models
Avg (78.0%)

BBH

Current model
Other models
Avg (81.1%)

IFEval

Current model
Other models
Avg (84.1%)

Providers Pricing Coming Soon

We're working on gathering comprehensive pricing data from all major providers for Olmo 2 32B. Compare costs across platforms to find the best pricing for your use case.

OpenAI
Anthropic
Google
Mistral AI
Cohere

Share your feedback

Hi, I'm Charlie Palars, the founder of Deepranking.ai. I'm always looking for ways to improve the site and make it more useful for you. You can write me through this form or directly through X at @palarsio.

Your feedback helps us improve our service

Stay Ahead with AI Updates

Get insights on Gemini Pro 2.5, Sonnet 3.7 and more top AI models