Command A logo

Command A

New
Released in the last 30 days

Cohere

Cohere’s Command A (Mar 2025) is a 111B parameter model that delivers GPT-4-tier performance with dramatically better efficiency—requiring just two A100s or H100s to deploy, compared to 8–32 for most competitors. It supports a 256K context window (2× GPT-4o), hybrid attention (sliding + global), and excels across enterprise-critical benchmarks: outperforming or matching GPT-4o and DeepSeek-V3 in human evals for business, coding, and agentic tasks. Notably, it leads on BFCL-v3 and Taubench for tool-using agents, performs on par with GPT-4o on MMLU, MBPPPlus, and SQL, and dominates RepoQA in long-context code understanding. With token throughput up to 156/sec—1.75× GPT-4o, 2.4× DeepSeek-V3—Command A is fast, accurate, and enterprise-ready. But what makes it useful is how well it translates benchmark wins into real-world capabilities. Command A is designed for RAG with citations, structured tool use, and multilingual fluency across 23 languages, including best-in-class dialect control (e.g., Arabic ADI2 scores). It ships with strict/contextual safety modes, runs well in private deployments, and integrates seamlessly into Cohere’s North agent stack for secure internal automation. Its open weights (non-commercial) and cost-effective footprint make it uniquely viable for teams building internal copilots or vertical agents. Among frontier-class LLMs, Command A may not shout the loudest—but it speaks the clearest, and fastest, where it counts.

Model Specifications

Technical details and capabilities of Command A

Core Specifications

111.0B Parameters

Model size and complexity

256.0K / 256.0K

Input / Output tokens

March 12, 2025

Last 30 Days

Release date

Capabilities & License

Multimodal Support
Not Supported
Web Hydrated
No
License
CC-BY-NC

Resources

API Reference
https://docs.cohere.com/reference/about
Playground
https://dashboard.cohere.com/welcome/login?redirect_uri=%2Fplayground%2Fchat%3Fmodel%3Dcommand-a-03-2025&ref=

Performance Insights

Check out how Command A handles various AI tasks through comprehensive benchmark results.

100
75
50
25
0
91
RepoQA 32k
91
(91%)
90
IFEval
90
(90%)
89
MBPPPlus
89
(89%)
84
MMLU
84
(84%)
78
MATH
78
(78%)
71
SQL
71
(71%)
65
BFCL
65
(65%)
60
Taubench Retail
60
(60%)
43
Taubench Airline
43
(43%)
RepoQA 32k
IFEval
MBPPPlus
MMLU
MATH
SQL
BFCL
Taubench Retail
Taubench Airline

Model Comparison

See how Command A stacks up against other leading models across key performance metrics.

100
80
60
40
20
0
84
MMLU - Command A
84
(84%)
85.9
MMLU - Nova Pro
85.9
(86%)
80.5
MMLU - Nova Lite
80.5
(81%)
88.7
MMLU - GPT-4o
88.7
(89%)
77.6
MMLU - Nova Micro
77.6
(78%)
88.5
MMLU - DeepSeek-V3
88.5
(89%)
78
MATH - Command A
78
(78%)
76.6
MATH - Nova Pro
76.6
(77%)
73.3
MATH - Nova Lite
73.3
(73%)
76.6
MATH - GPT-4o
76.6
(77%)
69.3
MATH - Nova Micro
69.3
(69%)
61.6
MATH - DeepSeek-V3
61.6
(62%)
90
IFEval - Command A
90
(90%)
92.1
IFEval - Nova Pro
92.1
(92%)
89.7
IFEval - Nova Lite
89.7
(90%)
84
IFEval - GPT-4o
84
(84%)
87.2
IFEval - Nova Micro
87.2
(87%)
86.1
IFEval - DeepSeek-V3
86.1
(86%)
65
BFCL - Command A
65
(65%)
68.4
BFCL - Nova Pro
68.4
(68%)
66.6
BFCL - Nova Lite
66.6
(67%)
74
BFCL - GPT-4o
74
(74%)
56.2
BFCL - Nova Micro
56.2
(56%)
60
BFCL - DeepSeek-V3
60
(60%)
MMLU
MATH
IFEval
BFCL
Command A
Nova Pro
Nova Lite
GPT-4o
Nova Micro
DeepSeek-V3

Detailed Benchmarks

Dive deeper into Command A's performance across specific task categories. Expand each section to see detailed metrics and comparisons.

Knowledge

MMLU

Current model
Other models
Avg (82.8%)

MATH

Current model
Other models
Avg (76.2%)

Non categorized

IFEval

Current model
Other models
Avg (87.2%)

Taubench Retail

Current model
Other models
Avg (58.3%)

Taubench Airline

Current model
Other models
Avg (38.0%)

BFCL

Current model
Other models
Avg (68.1%)

MBPPPlus

Current model
Other models
Avg (89.3%)

SQL

Current model
Other models
Avg (66.3%)

RepoQA 32k

Current model
Other models
Avg (89.0%)

Providers Pricing Coming Soon

We're working on gathering comprehensive pricing data from all major providers for Command A. Compare costs across platforms to find the best pricing for your use case.

OpenAI
Anthropic
Google
Mistral AI
Cohere

Share your feedback

Hi, I'm Charlie Palars, the founder of Deepranking.ai. I'm always looking for ways to improve the site and make it more useful for you. You can write me through this form or directly through X at @palarsio.

Your feedback helps us improve our service

Stay Ahead with AI Updates

Get insights on Gemini Pro 2.5, Sonnet 3.7 and more top AI models