DeepSeek: R1 Distill Qwen 32B

DeepSeek: R1 Distill Qwen 32B

deepseek · Released Jan 29, 2025
40
Our Score

DeepSeek R1 Distill Qwen 32B is a distilled large language model based on Qwen 2.5 32B, using outputs from DeepSeek R1. It outperforms OpenAI's o1-mini across various benchmarks, achieving new state-of-the-art results for dense models.nnOther benchmark results include:nn- AIME 2024 pass@1: 72.6n- MATH-500 pass@1: 94.3n- CodeForces Rating: 1691nnThe model leverages fine-tuning from DeepSeek R1's outputs, enabling competitive performance comparable to larger frontier models.

$0.29 / 1M Input Price
$0.29 / 1M Output Price
32,768 tokens Context Window
32,768 tokens Max Output
32B Parameters

Architecture

ModalityText → Text
TokenizerQwen
Instruct Typedeepseek-r1
Parameters32B

Performance Indices

Source: Artificial Analysis

17.2 Intelligence Index
63 Math Index

Benchmark Scores

Evaluations

GPQA Diamond 61.5%
Graduate-level scientific reasoning
HLE 5.5%
Humanity's Last Exam
MMLU Pro 73.9%
Multi-task language understanding
LiveCodeBench 27%
Live coding evaluation
SciCode 37.6%
Scientific computing
MATH 500 94.1%
Mathematical problem-solving
AIME 68.7%
Competition mathematics
AIME 2025 63%
Competition mathematics (2025)
IFBench 22.9%
Instruction following
LCR 9.7%
Long-context reasoning

Benchmark data from Artificial Analysis and Hugging Face

Model Information

OpenRouter ID deepseek/deepseek-r1-distill-qwen-32b
Providerdeepseek
Model FamilyDeepSeek
Release Date January 29, 2025
Context Length32,768 tokens
Max Completion32,768 tokens
Status Active

Pricing

Token Type Cost per 1M tokens Cost per 1K tokens
Input $0.29 $0.000290
Output $0.29 $0.000290

Live Performance

Live endpoint metrics — refreshed every 30 minutes.

99.8%
Avg Uptime
373ms
Best Latency (TTFT)
37 tok/s
Best Throughput
1/2
Active Endpoints
Available via: NextBit, Cloudflare