Anthropic: Claude Opus 4.1

Anthropic: Claude Opus 4.1

anthropic · Released Aug 5, 2025
70
Our Score

Claude Opus 4.1 is an updated version of Anthropic’s flagship model, offering improved performance in coding, reasoning, and agentic tasks. It achieves 74.5% on SWE-bench Verified and shows notable gains in multi-file code refactoring, debugging precision, and detail-oriented reasoning. The model supports extended thinking up to 64K tokens and is optimized for tasks involving research, data analysis, and tool-assisted reasoning.

$15.00 / 1M Input Price
$75.00 / 1M Output Price
200,000 tokens Context Window
32,000 tokens Max Output

Capabilities

Tool Use Function Calling Vision

Architecture

ModalityText + Image + File → Text
TokenizerClaude

Performance Indices

Source: Artificial Analysis

31.9 Intelligence Index
36.5 Coding Index
52.8 Agentic Index
80.3 Math Index

Benchmark Scores

Evaluations

GPQA Diamond 80.9%
Graduate-level scientific reasoning
HLE 11.9%
Humanity's Last Exam
MMLU Pro 88%
Multi-task language understanding
LiveCodeBench 65.4%
Live coding evaluation
SciCode 40.9%
Scientific computing
AIME 2025 80.3%
Competition mathematics (2025)
IFBench 55.4%
Instruction following
LCR 66.3%
Long-context reasoning
TerminalBench Hard 34.3%
Agentic terminal tasks
τ²-Bench 71.4%
Conversational agent benchmark

Benchmark data from Artificial Analysis and Hugging Face

Model Information

OpenRouter ID anthropic/claude-opus-4.1
Provideranthropic
Release Date August 5, 2025
Context Length200,000 tokens
Max Completion32,000 tokens
Status Active

Pricing

Token Type Cost per 1M tokens Cost per 1K tokens
Input $15.00 $0.015000
Output $75.00 $0.075000

Live Performance

Live endpoint metrics — refreshed every 30 minutes.

100%
Avg Uptime
1,532ms
Best Latency (TTFT)
21 tok/s
Best Throughput
1/5
Active Endpoints
Available via: Amazon Bedrock, Google, Anthropic

Leaderboard Categories