Qwen: Qwen3 235B A22B Instruct 2507

Qwen: Qwen3 235B A22B Instruct 2507

qwen · Released Jul 21, 2025
38
Our Score

Qwen3-235B-A22B-Instruct-2507 is a multilingual, instruction-tuned mixture-of-experts language model based on the Qwen3-235B architecture, with 22B active parameters per forward pass. It is optimized for general-purpose text generation, including instruction following, logical reasoning, math, code, and tool usage. The model supports a native 262K context length and does not implement "thinking mode" ( blocks). Compared to its base variant, this version delivers significant gains in knowledge coverage, long-context reasoning, coding benchmarks, and alignment with open-ended tasks. It is particularly strong on multilingual understanding, math reasoning (e.g., AIME, HMMT), and alignment evaluations like Arena-Hard and WritingBench.

$0.07 / 1M Input Price
$0.10 / 1M Output Price
262,144 tokens Context Window
235B Parameters

Capabilities

Tool Use Function Calling

Architecture

ModalityText → Text
TokenizerQwen3
Parameters235B

Model Information

OpenRouter ID qwen/qwen3-235b-a22b-2507
Providerqwen
Release Date July 21, 2025
Context Length262,144 tokens
Status Active

Pricing

Token Type Cost per 1M tokens Cost per 1K tokens
Input $0.07 $0.000071
Output $0.10 $0.000100

Live Performance

Live endpoint metrics — refreshed every 30 minutes.

96.3%
Avg Uptime
252ms
Best Latency (TTFT)
64 tok/s
Best Throughput
12/14
Active Endpoints
Available via: DeepInfra, Chutes, Novita, SiliconFlow, WandB, Parasail, Alibaba, Together +5 more

Leaderboard Categories