Qwen: Qwen3 235B A22B Thinking 2507

Qwen: Qwen3 235B A22B Thinking 2507

qwen · Released Jul 25, 2025
35
Our Score

Qwen3-235B-A22B-Thinking-2507 is a high-performance, open-weight Mixture-of-Experts (MoE) language model optimized for complex reasoning tasks. It activates 22B of its 235B parameters per forward pass and natively supports up to 262,144 tokens of context. This "thinking-only" variant enhances structured logical reasoning, mathematics, science, and long-form generation, showing strong benchmark performance across AIME, SuperGPQA, LiveCodeBench, and MMLU-Redux. It enforces a special reasoning mode () and is designed for high-token outputs (up to 81,920 tokens) in challenging domains. The model is instruction-tuned and excels at step-by-step reasoning, tool use, agentic workflows, and multilingual tasks. This release represents the most capable open-source variant in the Qwen3-235B series, surpassing many closed models in structured reasoning use cases.

$0.11 / 1M Input Price
$0.60 / 1M Output Price
262,144 tokens Context Window
262,144 tokens Max Output
235B Parameters

Capabilities

Tool Use Function Calling

Architecture

ModalityText → Text
TokenizerQwen3
Instruct Typeqwen3
Parameters235B

Model Information

OpenRouter ID qwen/qwen3-235b-a22b-thinking-2507
Providerqwen
Release Date July 25, 2025
Context Length262,144 tokens
Max Completion262,144 tokens
Status Active

Pricing

Token Type Cost per 1M tokens Cost per 1K tokens
Input $0.11 $0.000110
Output $0.60 $0.000600

Live Performance

Live endpoint metrics — refreshed every 30 minutes.

100%
Avg Uptime
588ms
Best Latency (TTFT)
82 tok/s
Best Throughput
2/5
Active Endpoints
Available via: Chutes, Alibaba, DeepInfra, AtlasCloud, Novita

Leaderboard Categories