
Claude 3.7 Sonnet
Anthropic
Claude 3.7 Sonnet is Anthropic’s most advanced model to date and the first hybrid reasoning model on the market, blending near-instant responses with a new "extended thinking" mode that visibly improves performance on complex reasoning tasks. This dual-mode capability allows users to trade speed for depth by adjusting a thinking token budget—up to 128K output tokens—enabling stronger results in math, physics, instruction-following, and software planning. Unlike prior models that silo reasoning into separate modes or products, Sonnet unifies them under a single API, creating a seamless experience for developers and end users alike. It's especially strong in enterprise-facing use cases, where Claude’s long context window, low hallucination rate, and agentic planning make it well-suited for content analysis, customer support, and data extraction. Where Claude 3.7 Sonnet truly stands out is in software engineering. It achieves **state-of-the-art results on SWE-bench Verified**—scoring 63.7% pass@1 with minimal scaffolding and 70.3% with higher test-time compute—surpassing GPT-4 Turbo and DeepSeek-VL. It also leads on TAU-bench, a framework for real-world agent tasks, reflecting Claude’s strength in multi-step tool use and planning. Claude Code, a new CLI-based coding agent, extends this further, completing end-to-end dev tasks in a single pass and outperforming rivals in handling complex codebases, according to testing by Cursor, Cognition, Replit, and Canva. With deeply integrated GitHub tooling and superior design intuition, Claude 3.7 is not only the best Claude model—it’s arguably the most practically useful LLM for coding and real-world reasoning workflows available today.
Model Specifications
Technical details and capabilities of Claude 3.7 Sonnet
Core Specifications
200.0K / 128.0K
Input / Output tokens
September 30, 2024
Knowledge cutoff date
February 23, 2025
Release date
Performance Insights
Check out how Claude 3.7 Sonnet handles various AI tasks through comprehensive benchmark results.
Model Comparison
See how Claude 3.7 Sonnet stacks up against other leading models across key performance metrics.
Detailed Benchmarks
Dive deeper into Claude 3.7 Sonnet's performance across specific task categories. Expand each section to see detailed metrics and comparisons.
Math
MATH-500
AIME 2024
AIME 2025
Coding
SWE-bench Verified
Aider Polyglot
Knowledge
GPQA
Non categorized
TAU-bench Retail
TAU-bench Airline
MMMU
IFEval
Humanity's Last Exam
Providers Pricing Coming Soon
We're working on gathering comprehensive pricing data from all major providers for Claude 3.7 Sonnet. Compare costs across platforms to find the best pricing for your use case.
Share your feedback
Hi, I'm Charlie Palars, the founder of Deepranking.ai. I'm always looking for ways to improve the site and make it more useful for you. You can write me through this form or directly through X at @palarsio.
Your feedback helps us improve our service