Z.ai: GLM 4.6

Z.ai: GLM 4.6

z-ai · Released Sep 30, 2025 Specialist
62.6
Our Score

Performance Profile

Intelligence6Technical5.9Value7.5Content5.5
Intelligence 6/10
Technical 5.9/10
Content 5.5/10
Value 7.5/10

Compared with GLM-4.5, this generation brings several key improvements: Longer context window: The context window has been expanded from 128K to 200K tokens, enabling the model to handle more complex..

$0.39 / 1M
Input Price
$1.90 / 1M
Output Price
204,800 tokens
Context Window
204,800 tokens
Max Output

Capabilities

Tool Use Function Calling

Architecture

ModalityText → Text
TokenizerOther

Performance Indices

Source: Artificial Analysis

32.5 Intelligence Index
29.5 Coding Index
47.8 Agentic Index
86 Math Index

Benchmark Scores

Intelligence

GPQA Diamond 78% Graduate-level scientific reasoning
HLE 13.3% Humanity's Last Exam
MMLU Pro 82.9% Multi-task language understanding
AIME 2025 86% Competition mathematics (2025)
SciCode 38.4% Scientific computing

Technical

LiveCodeBench 69.5% Live coding evaluation
TerminalBench Hard 25% Agentic terminal tasks
τ²-Bench 70.5% Conversational agent benchmark

Content

IFBench 43.4% Instruction following
LCR 54.3% Long-context reasoning

Benchmark data from Artificial Analysis and Hugging Face

How does Z.ai: GLM 4.6 stack up?

Compare side-by-side with other specialist models.

Compare Models

Model Information

OpenRouter ID z-ai/glm-4.6
Providerz-ai
Release Date September 30, 2025
Context Length204,800 tokens
Max Completion204,800 tokens
Status Active

Pricing

Token Type Cost per 1M tokens Cost per 1K tokens
Input $0.39 $0.000390
Output $1.90 $0.001900

Live Performance

Live endpoint metrics — refreshed every 30 minutes.

98.8%
Avg Uptime
893ms
Best Latency (TTFT)
44 tok/s
Best Throughput
5/6
Active Endpoints
Available via: SiliconFlow, DeepInfra, AtlasCloud, Novita, BaseTen, Z.AI

Leaderboard Categories