Magistral Small is a compact, high-performance language model optimized for efficient inference while maintaining strong capabilities across various tasks.
Added Sep 25, 2025
Context Window
32.8K
Max Output
32.8K
Input Price (Auto)
$0.40/1M
Output Price (Auto)
$1.40/1M
Performance metrics and benchmarks
Sourced from Artificial Analysis.
Coding Index
11.1
Auto routing is available for this model. Explicit provider selection is not available.
Loading provider options…
GPQA Diamond
Graduate-level scientific reasoning
64.1%
Better than 53% of models compared
HLE
Humanity's Last Exam
7.2%
Better than 62% of models compared
IFBench
Instruction-following benchmark
24.8%
Better than 9% of models compared
T²-Bench Telecom
Conversational AI agents in dual-control scenarios
26.6%
Better than 37% of models compared
AA-LCR
Long context reasoning evaluation
0.0%
Better than 7% of models compared
CritPt
Research-level physics reasoning
0.0%
Better than 36% of models compared
SciCode
Python programming for scientific computing
24.1%
Better than 35% of models compared
Terminal-Bench Hard
Agentic coding and terminal use
4.5%
Better than 31% of models compared
AA-Omniscience Accuracy
Proportion of correctly answered questions
18.3%
Better than 53% of models compared
AA-Omniscience Hallucination Rate
Rate of incorrect answers among non-correct responses
75.5%
Better than 70% of models compared
Last updated May 15, 2026, 8:34 PM
Artificial Analysis