Anthropic's updated faster and cheaper model, offering good results on chatbots and coding.
Context Window
200.0K
Max Output
8.2K
Input Price (Auto)
$0.80/1M
Output Price (Auto)
$4.00/1M
Cache Read (Auto)
$0.080/1M
Capabilities
Performance metrics and benchmarks
Sourced from Artificial Analysis.
Intelligence Index
18.7
Auto routing is available for this model. Explicit provider selection is not available.
Loading provider options…
Coding Index
10.7
GPQA Diamond
Graduate-level scientific reasoning
40.8%
Better than 22% of models compared
HLE
Humanity's Last Exam
3.5%
Better than 5% of models compared
IFBench
Instruction-following benchmark
42.8%
Better than 54% of models compared
T²-Bench Telecom
Conversational AI agents in dual-control scenarios
24.6%
Better than 32% of models compared
AA-LCR
Long context reasoning evaluation
23.3%
Better than 42% of models compared
SciCode
Python programming for scientific computing
27.4%
Better than 43% of models compared
Terminal-Bench Hard
Agentic coding and terminal use
2.3%
Better than 22% of models compared
AIME
American Invitational Mathematics Examination
3.3%
Better than 15% of models compared
Math-500
Diverse mathematical problem solving benchmark
72.1%
Better than 33% of models compared
MMLU-Pro
Professional and academic subject knowledge
63.4%
Better than 25% of models compared
Last updated May 15, 2026
Artificial AnalysisLiveCodeBench
Contamination-free coding benchmark
31.4%
Better than 37% of models compared