Qwen: Qwen3 14B

Qwen: Qwen3 14B

qwen · Released Apr 28, 2025
33
Our Score

Qwen3-14B is a dense 14.8B parameter causal language model from the Qwen3 series, designed for both complex reasoning and efficient dialogue. It supports seamless switching between a "thinking" mode for tasks like math, programming, and logical inference, and a "non-thinking" mode for general-purpose conversation. The model is fine-tuned for instruction-following, agent tool use, creative writing, and multilingual tasks across 100+ languages and dialects. It natively handles 32K token contexts and can extend to 131K tokens using YaRN-based scaling.

$0.06 / 1M Input Price
$0.24 / 1M Output Price
40,960 tokens Context Window
40,960 tokens Max Output
14B Parameters

Capabilities

Tool Use Function Calling

Architecture

ModalityText → Text
TokenizerQwen3
Instruct Typeqwen3
Parameters14B

Performance Indices

Source: Artificial Analysis

16.2 Intelligence Index
13.1 Coding Index
19.2 Agentic Index
55.7 Math Index

Benchmark Scores

Evaluations

GPQA Diamond 60.4%
Graduate-level scientific reasoning
HLE 4.3%
Humanity's Last Exam
MMLU Pro 77.4%
Multi-task language understanding
LiveCodeBench 52.3%
Live coding evaluation
SciCode 31.6%
Scientific computing
MATH 500 96.1%
Mathematical problem-solving
AIME 76.3%
Competition mathematics
AIME 2025 55.7%
Competition mathematics (2025)
IFBench 40.5%
Instruction following
TerminalBench Hard 3.8%
Agentic terminal tasks
τ²-Bench 34.5%
Conversational agent benchmark

Benchmark data from Artificial Analysis and Hugging Face

Model Information

OpenRouter ID qwen/qwen3-14b
Providerqwen
Release Date April 28, 2025
Context Length40,960 tokens
Max Completion40,960 tokens
Status Active

Pricing

Token Type Cost per 1M tokens Cost per 1K tokens
Input $0.06 $0.000060
Output $0.24 $0.000240

Live Performance

Live endpoint metrics — refreshed every 30 minutes.

100%
Avg Uptime
346ms
Best Latency (TTFT)
54 tok/s
Best Throughput
1/3
Active Endpoints
Available via: NextBit, DeepInfra, Alibaba