Anthropic: Claude Opus 4

Anthropic: Claude Opus 4

anthropic · Released May 22, 2025 Professional
Intelligence #41 / 525
74.9 Our Score
Speed #217 / 244
37.3 tokens / sec
Input #514 / 527
$15.00 per 1M tokens
Output #518 / 527
$75.00 per 1M tokens
Context #151 / 527
200,000 tokens

Analysis Summary

Anthropic: Claude Opus 4 sits in the Professional tier on our leaderboard, ranked #41 of 525 published models on overall intelligence. At $15.00 input and $75.00 output per 1M tokens, it is among the most expensive on the market. It offers a generous context window for extended reasoning and code review and supports tool use, function calling, vision, and reasoning.

Editorial notes

Claude Opus 4 from Anthropic is a highly capable model with strong reasoning, solid coding performance, vision support, and excellent instruction-following — a reliable choice for demanding content and agentic business tasks, though its premium pricing limits value for high-volume use.

Assessed April 16, 2026

Rankings consider pricing, capabilities, benchmarks, and real-world applicability and are refreshed as new models launch. Feedback?

Performance Profile

Intelligence6.9Technical6.5Value5Content8
Intelligence 6.9/10
Technical 6.5/10
Content 8/10
Value 5/10

Claude Opus 4 is benchmarked as the world’s best coding model, at time of release, bringing sustained performance on complex, long-running tasks and agent workflows. It sets new benchmarks in..

Capabilities

Tool Use Function Calling Vision

Performance Indices

Source: Artificial Analysis

39 Intelligence Index
34 Coding Index
52.3 Agentic Index
73.3 Math Index

Benchmark Scores

Intelligence

GPQA Diamond 79.6% Graduate-level scientific reasoning
HLE 11.7% Humanity's Last Exam
MMLU Pro 87.3% Multi-task language understanding
MATH 500 98.2% Mathematical problem-solving
AIME 75.7% Competition mathematics
AIME 2025 73.3% Competition mathematics (2025)
SciCode 39.8% Scientific computing

Technical

LiveCodeBench 63.6% Live coding evaluation
TerminalBench Hard 31.1% Agentic terminal tasks
τ²-Bench 73.4% Conversational agent benchmark

Content

IFBench 53.7% Instruction following
LCR 33.7% Long-context reasoning

Benchmark data from Artificial Analysis and Hugging Face

How does Anthropic: Claude Opus 4 stack up?

Compare side-by-side with other professional models.

Compare Models

Model Information

OpenRouter ID anthropic/claude-opus-4
Provideranthropic
Release Date May 22, 2025
Context Length200,000 tokens
Max Completion32,000 tokens
Status Active

Pricing

Token Type Cost per 1M tokens Cost per 1K tokens
Input $15.00 $0.015000
Output $75.00 $0.075000

Live Performance

Live endpoint metrics — refreshed every 30 minutes.

99.6%
Avg Uptime
1,456ms
Best Latency (TTFT)
23 tok/s
Best Throughput
1/4
Active Endpoints
Available via: Google, Anthropic, Amazon Bedrock

Leaderboard Categories