Discover AI language models for conversations, coding, and creative writing
GLM 5 Turbo
Fast OpenRouter-hosted GLM 5 Turbo variant from Z-AI for general chat, coding, and tool use. Not included in the subscription.
Benchmarks (Artificial Analysis)
Intelligence
49.8
Coding
44.2
Speed
66.5
Features
Context
202.8K
Max Output
131.1K
Date Added
Mar 15, 2026
Pricing
Input:
$0.96/1M
Output:
$3.20/1M
Est./msg:
$0.0026
Qwen3.5 27B TEE
Qwen3.5 27B is a native vision-language dense model optimized for fast responses while balancing quality and inference speed. Running inside a TEE (Trusted Execution Environment), with verifiably no logging by the provider.
Benchmarks (Artificial Analysis)
Intelligence
40.1
Coding
37.4
Speed
58.4
Features
Context
262.1K
Max Output
65.5K
Date Added
Mar 13, 2026
Pricing
Input:
$0.30/1M
Output:
$2.40/1M
Est./msg:
$0.0015
Grok 4.20 Beta Reasoning
Grok 4.20 Beta with reasoning enabled. xAI positions it as a flagship model with fast agentic tool calling, strong prompt adherence, and a 2M-token context window.
Features
Context
2.0M
Max Output
131.1K
Date Added
Mar 12, 2026
Pricing
Input:
$2.00/1M
Output:
$6.00/1M
Est./msg:
$0.0050
Grok 4.20 Beta Non-Reasoning
Grok 4.20 Beta with reasoning disabled for lower-latency tool-centric workloads, while keeping the same 2M-token context window and multimodal input support.
Features
Context
2.0M
Max Output
131.1K
Date Added
Mar 12, 2026
Pricing
Input:
$2.00/1M
Output:
$6.00/1M
Est./msg:
$0.0050
Grok 4.20 Multi-Agent Beta
Grok 4.20 Multi-Agent Beta tuned for agentic workflows with tool calling, structured outputs, reasoning support, and a 2M-token context window.
Benchmarks (Artificial Analysis)
Intelligence
13.3
Features
Context
2.0M
Max Output
131.1K
Date Added
Mar 12, 2026
Pricing
Input:
$2.00/1M
Output:
$6.00/1M
Est./msg:
$0.0050
Nvidia Nemotron 3 Super 120B
Nvidia's Nemotron 3 Super 120B A12B model from the March 2026 Nemotron 3 release. It uses a hybrid Mamba-Transformer MoE architecture and targets agentic and coding workloads with a 262K context window here.
Benchmarks (Artificial Analysis)
Intelligence
36.0
Coding
31.2
Speed
433.5
Features
Context
262.1K
Max Output
16.4K
Date Added
Mar 11, 2026
Pricing
Input:
$0.05/1M
Output:
$0.25/1M
Est./msg:
$0.0002
Nvidia Nemotron 3 Super 120B Thinking
Nvidia Nemotron 3 Super 120B with reasoning content enabled. Returns separate thinking content when requested.
Features
Context
262.1K
Max Output
16.4K
Date Added
Mar 11, 2026
Pricing
Input:
$0.05/1M
Output:
$0.25/1M
Est./msg:
$0.0002
ByteDance Seed 2.0 Lite
ByteDance Seed 2.0 Lite is a balanced long-context model for high-frequency enterprise workloads, tuned for unstructured information processing, text creation, search and recommendation, and stable structured outputs. Supports 262k context on OpenRouter and routes via BytePlus Singapore, which claims no logging.
Features
Context
262.1K
Max Output
131.1K
Date Added
Mar 10, 2026
Pricing
Input:
$0.25/1M
Output:
$2.00/1M
Est./msg:
$0.0013
GPT 5.4
GPT-5.4 is OpenAI's latest frontier model for professional work with stronger reasoning, coding, and tool use.
Benchmarks (Artificial Analysis)
Intelligence
57.2
Coding
57.3
Speed
85.4
Features
Context
922.0K
Max Output
128.0K
Date Added
Mar 5, 2026
Pricing
Input:
$2.50/1M
Output:
$15.00/1M
Cache:
Read $0.25/1M
Est./msg:
$0.0100
GPT 5.4 Pro
The highest-performing GPT-5.4 variant from OpenAI, tuned for complex reasoning and demanding professional tasks.
Features
Context
922.0K
Max Output
128.0K
Date Added
Mar 5, 2026
Pricing
Input:
$30.00/1M
Output:
$180.00/1M
Cache:
Read $3.00/1M
Est./msg:
$0.1200
GPT 5.3 Chat
Chat-optimized GPT-5.3 variant with improved conversational quality and instruction following.
Benchmarks (Artificial Analysis)
Intelligence
54.0
Coding
53.1
Speed
69.5
Features
Context
128.0K
Max Output
16.4K
Date Added
Mar 3, 2026
Pricing
Input:
$1.75/1M
Output:
$14.00/1M
Cache:
Read $0.18/1M
Est./msg:
$0.0088
Gemini 3.1 Flash Lite Preview
Google's Gemini 3.1 Flash Lite Preview via OpenRouter. Optimized for high-volume, cost-efficient reasoning and multimodal workloads.
Benchmarks (Artificial Analysis)
Intelligence
33.5
Coding
30.1
Speed
274.2
Features
Context
1.0M
Max Output
65.5K
Date Added
Mar 3, 2026
Pricing
Input:
$0.25/1M
Output:
$1.50/1M
Cache:
Read $0.03/1M
Est./msg:
$0.0010
Solar Pro 3
Upstage's Solar Pro 3 is a Mixture-of-Experts (MoE) language model with 102B total parameters and 12B active parameters per forward pass, optimized for Korean with strong English and Japanese support.
Context
128.0K
Max Output
128.0K
Date Added
Mar 3, 2026
Pricing
Input:
$0.15/1M
Output:
$0.60/1M
Est./msg:
$0.0004
Qwen3.5 397B A17B TEE
Qwen 3.5's open-source 397B MoE model (17B active params) with hybrid linear attention. Running inside a TEE (Trusted Execution Environment), with verifiably no logging by the provider.
Benchmarks (Artificial Analysis)
Intelligence
40.1
Coding
37.4
Speed
58.4
Context
258.0K
Max Output
65.5K
Date Added
Feb 28, 2026
Pricing
Input:
$0.60/1M
Output:
$3.60/1M
Est./msg:
$0.0024
Gemini 3.1 Pro (Preview Custom Tools)
Gemini 3.1 Pro preview variant tuned for better tool selection behavior in coding agents and multi-tool workflows. It reduces overuse of generic bash tools and improves function-calling reliability while retaining Gemini 3.1 Pro's multimodal reasoning and 1M-token context. NOTE: Inputs > 200k tokens are charged at 2x input and 1.5x output rates.
Benchmarks (Artificial Analysis)
Intelligence
57.2
Coding
55.5
Speed
120.4
Features
Context
1.0M
Max Output
65.5K
Date Added
Feb 27, 2026
Pricing
Input:
$2.00/1M
Output:
$12.00/1M
Cache:
Read $0.20/1M
Est./msg:
$0.0080
LFM2 24B A2B
A 24B-parameter Mixture-of-Experts model from LiquidAI with 2B active parameters per token, designed for efficient high-quality generation and practical local deployment.
Benchmarks (Artificial Analysis)
Intelligence
10.5
Coding
3.6
Speed
241.3
Context
32.8K
Max Output
32.8K
Date Added
Feb 25, 2026
Pricing
Input:
$0.03/1M
Output:
$0.12/1M
Est./msg:
$0.0001
GPT 5.3 Codex
Coding-focused GPT-5.3 variant with optimized routing.
Benchmarks (Artificial Analysis)
Intelligence
54.0
Coding
53.1
Speed
69.5
Features
Context
400.0K
Max Output
128.0K
Date Added
Feb 24, 2026
Pricing
Input:
$1.75/1M
Output:
$14.00/1M
Cache:
Read $0.18/1M
Est./msg:
$0.0088
Qwen3.5 122B A10B
Qwen3.5 122B A10B is Alibaba's high-end native vision-language model in the Qwen 3.5 family with strong text and multimodal performance.
Benchmarks (Artificial Analysis)
Intelligence
40.1
Coding
37.4
Speed
58.4
Features
Context
260.1K
Max Output
65.5K
Date Added
Feb 24, 2026
Pricing
Input:
$0.36/1M
Output:
$2.88/1M
Est./msg:
$0.0018
View Providers
Qwen3.5 122B A10B Thinking
Qwen3.5 122B A10B with extended reasoning enabled. Alibaba's high-end native vision-language model in the Qwen 3.5 family.
Benchmarks (Artificial Analysis)
Intelligence
39.9
Coding
30.5
Speed
35.5
Features
Context
260.1K
Max Output
65.5K
Date Added
Feb 24, 2026
Pricing
Input:
$0.36/1M
Output:
$2.88/1M
Est./msg:
$0.0018
View Providers
Qwen3.5 27B
Qwen3.5 27B is a native vision-language dense model optimized for fast responses while balancing quality and inference speed.
Benchmarks (Artificial Analysis)
Intelligence
40.1
Coding
37.4
Speed
58.4
Features
Context
260.1K
Max Output
65.5K
Date Added
Feb 24, 2026
Pricing
Input:
$0.27/1M
Output:
$2.16/1M
Est./msg:
$0.0014
View Providers
Qwen3.5 27B Thinking
Qwen3.5 27B with extended reasoning enabled. A native vision-language dense model optimized for fast responses.
Benchmarks (Artificial Analysis)
Intelligence
39.9
Coding
30.5
Speed
35.5
Features
Context
260.1K
Max Output
65.5K
Date Added
Feb 24, 2026
Pricing
Input:
$0.27/1M
Output:
$2.16/1M
Est./msg:
$0.0014
View Providers
Qwen3.5 35B A3B
Qwen3.5 35B A3B is a native vision-language MoE model with hybrid attention designed for efficient inference and strong general performance.
Benchmarks (Artificial Analysis)
Intelligence
40.1
Coding
37.4
Speed
58.4
Features
Context
260.1K
Max Output
65.5K
Date Added
Feb 24, 2026
Pricing
Input:
$0.22/1M
Output:
$1.80/1M
Est./msg:
$0.0011
View Providers
Qwen3.5 35B A3B Thinking
Qwen3.5 35B A3B with extended reasoning enabled. A native vision-language MoE model with hybrid attention.
Benchmarks (Artificial Analysis)
Intelligence
39.9
Coding
30.5
Speed
35.5
Features
Context
260.1K
Max Output
65.5K
Date Added
Feb 24, 2026
Pricing
Input:
$0.22/1M
Output:
$1.80/1M
Est./msg:
$0.0011
View Providers
Qwen3.5 Flash
Qwen3.5 Flash is the fastest and cheapest native Qwen 3.5 vision-language model with a 1M-token context window.
Benchmarks (Artificial Analysis)
Intelligence
40.1
Coding
37.4
Speed
58.4
Features
Context
991.8K
Max Output
65.5K
Date Added
Feb 24, 2026
Pricing
Input:
$0.09/1M
Output:
$0.36/1M
Est./msg:
$0.0003
Qwen3.5 Flash Thinking
Qwen3.5 Flash with extended reasoning enabled. The fastest and cheapest native Qwen 3.5 vision-language model.
Benchmarks (Artificial Analysis)
Intelligence
39.9
Coding
30.5
Speed
35.5
Features
Context
991.8K
Max Output
65.5K
Date Added
Feb 24, 2026
Pricing
Input:
$0.09/1M
Output:
$0.36/1M
Est./msg:
$0.0003
AionLabs: Aion-2.0
Aion-2.0 is a DeepSeek V3.2 variant tuned for immersive roleplay and storytelling with stronger tension, conflict, and darker thematic nuance.
Context
131.1K
Max Output
32.8K
Date Added
Feb 23, 2026
Pricing
Input:
$1.36/1M
Output:
$2.72/1M
Est./msg:
$0.0027
Gemini 3.1 Pro (Preview High)
Gemini 3.1 Pro preview high-reasoning variant.
Benchmarks (Artificial Analysis)
Intelligence
57.2
Coding
55.5
Speed
120.4
Features
Context
1.0M
Max Output
65.5K
Date Added
Feb 21, 2026
Pricing
Input:
$2.00/1M
Output:
$12.00/1M
Cache:
Read $0.20/1M
Est./msg:
$0.0080
Gemini 3.1 Pro (Preview Low)
Gemini 3.1 Pro preview low-reasoning variant.
Benchmarks (Artificial Analysis)
Intelligence
57.2
Coding
55.5
Speed
120.4
Features
Context
1.0M
Max Output
65.5K
Date Added
Feb 21, 2026
Pricing
Input:
$2.00/1M
Output:
$12.00/1M
Cache:
Read $0.20/1M
Est./msg:
$0.0080
Gemini 3.1 Pro (Preview)
Gemini 3.1 Pro preview is built for tasks where simple answers are not enough. Stronger core reasoning for complex coding, math, and long-context workflows, with multimodal support and a reported 77.1% verified score on ARC-AGI-2. NOTE: Inputs > 200k tokens are charged at 2x input and 1.5x output rates.
Benchmarks (Artificial Analysis)
Intelligence
57.2
Coding
55.5
Speed
120.4
Features
Context
1.0M
Max Output
65.5K
Date Added
Feb 19, 2026
Pricing
Input:
$2.00/1M
Output:
$12.00/1M
Cache:
Read $0.20/1M
Est./msg:
$0.0080
Claude Sonnet 4.6
Claude Sonnet 4.6 is Anthropic's most capable Sonnet yet — a full upgrade across coding, computer use, long-context reasoning, agent planning, and design. Supports up to a 1M-token context window at Sonnet pricing.
Benchmarks (Artificial Analysis)
Intelligence
44.4
Coding
46.4
Speed
49.9
Features
Context
1.0M
Max Output
128.0K
Date Added
Feb 17, 2026
Pricing
Input:
$2.99/1M
Output:
$14.99/1M
Cache:
Read $0.30/1M · Write $3.74/1M (5m) / $5.98/1M (1h)
Est./msg:
$0.0105