Mixtral 8x22B is a powerful sparse Mixture of Experts (MoE) model with 141B total parameters and 39B active per token. Features a 64K context window, exceptional math performance, and cost-efficient inference. Supports English, French, German, Spanish, and Italian. Apache 2.0 licensed.
Added Dec 11, 2025
Context Window
65.5K
Max Output
32.8K
Input Price (Auto)
$0.90/1M
Output Price (Auto)
$0.90/1M
Performance metrics and benchmarks
Sourced from Artificial Analysis.
Intelligence Index
9.8
Auto routing is available for this model. Explicit provider selection is not available.
Loading provider options…
GPQA Diamond
Graduate-level scientific reasoning
33.2%
Better than 14% of models compared
HLE
Humanity's Last Exam
4.1%
Better than 18% of models compared
SciCode
Python programming for scientific computing
18.8%
Better than 23% of models compared
LiveCodeBench
Contamination-free coding benchmark
14.8%
Better than 15% of models compared
AIME
American Invitational Mathematics Examination
0.0%
Better than 3% of models compared
Math-500
Diverse mathematical problem solving benchmark
54.5%
Better than 17% of models compared
MMLU-Pro
Professional and academic subject knowledge
53.7%
Better than 17% of models compared
Last updated May 15, 2026
Artificial Analysis