Qwen: QwQ 32B

Qwen: QwQ 32B

qwen · Released Mar 5, 2025
55
Our Score

QwQ is the reasoning model of the Qwen series. Compared with conventional instruction-tuned models, QwQ, which is capable of thinking and reasoning, can achieve significantly enhanced performance in downstream tasks, especially hard problems. QwQ-32B is the medium-sized reasoning model, which is capable of achieving competitive performance against state-of-the-art reasoning models, e.g., DeepSeek-R1, o1-mini.

$0.15 / 1M Input Price
$0.40 / 1M Output Price
32,768 tokens Context Window
32,768 tokens Max Output
32B Parameters

Capabilities

Tool Use Function Calling

Architecture

ModalityText → Text
TokenizerQwen
Instruct Typeqwq
Parameters32B

Performance Indices

Source: Artificial Analysis

19.7 Intelligence Index
29 Math Index

Benchmark Scores

Evaluations

GPQA Diamond 59.3%
Graduate-level scientific reasoning
HLE 8.2%
Humanity's Last Exam
MMLU Pro 76.4%
Multi-task language understanding
LiveCodeBench 63.1%
Live coding evaluation
SciCode 35.8%
Scientific computing
MATH 500 95.7%
Mathematical problem-solving
AIME 78%
Competition mathematics
AIME 2025 29%
Competition mathematics (2025)
IFBench 38.8%
Instruction following
LCR 25%
Long-context reasoning

Benchmark data from Artificial Analysis and Hugging Face

Model Information

OpenRouter ID qwen/qwq-32b
Providerqwen
Release Date March 5, 2025
Context Length32,768 tokens
Max Completion32,768 tokens
Status Active

Pricing

Token Type Cost per 1M tokens Cost per 1K tokens
Input $0.15 $0.000150
Output $0.40 $0.000400

Live Performance

Live endpoint metrics — refreshed every 30 minutes.

1,301ms
Best Latency (TTFT)
60 tok/s
Best Throughput
0/2
Active Endpoints
Available via: NextBit, SiliconFlow

Leaderboard Categories