Based on the new Qwen3‑Next architecture (hybrid attention, highly sparse MoE, training‑stability optimizations, and multi‑token prediction), the Qwen3‑Next‑80B‑A3B‑Instruct model delivers extreme efficiency with only 3B active parameters per pass. It performs comparably to Qwen3‑235B‑A22B‑Instruct‑2507 and shows clear advantages on ultra‑long context tasks (up to 256K tokens).
Added Sep 11, 2025
Context Window
256.0K
Max Output
262.1K
Input Price (Auto)
$0.095/1M
Output Price (Auto)
$1.16/1M
Capabilities
Performance metrics and benchmarks
Sourced from Artificial Analysis.
Intelligence Index
20.1
Choose explicit providers for this model. Auto routing remains available as the default option.
Loading provider options…
Coding Index
15.3
GPQA Diamond
Graduate-level scientific reasoning
73.8%
Better than 69% of models compared
HLE
Humanity's Last Exam
7.3%
Better than 62% of models compared
IFBench
Instruction-following benchmark
39.7%
Better than 46% of models compared
T²-Bench Telecom
Conversational AI agents in dual-control scenarios
21.6%
Better than 27% of models compared
AA-LCR
Long context reasoning evaluation
51.3%
Better than 68% of models compared
SciCode
Python programming for scientific computing
30.7%
Better than 54% of models compared
Terminal-Bench Hard
Agentic coding and terminal use
7.6%
Better than 45% of models compared
AIME 2025
American Invitational Mathematics Examination 2025
66.3%
Better than 61% of models compared
MMLU-Pro
Professional and academic subject knowledge
81.9%
Better than 77% of models compared
Last updated May 15, 2026
Artificial AnalysisLiveCodeBench
Contamination-free coding benchmark
68.4%
Better than 76% of models compared