Meta: Llama 4 Maverick

Meta: Llama 4 Maverick

meta-llama · Released Apr 5, 2025
50
Our Score

Llama 4 Maverick 17B Instruct (128E) is a high-capacity multimodal language model from Meta, built on a mixture-of-experts (MoE) architecture with 128 experts and 17 billion active parameters per forward pass (400B total). It supports multilingual text and image input, and produces multilingual text and code output across 12 supported languages. Optimized for vision-language tasks, Maverick is instruction-tuned for assistant-like behavior, image reasoning, and general-purpose multimodal interaction. Maverick features early fusion for native multimodality and a 1 million token context window. It was trained on a curated mixture of public, licensed, and Meta-platform data, covering ~22 trillion tokens, with a knowledge cutoff in August 2024. Released on April 5, 2025 under the Llama 4 Community License, Maverick is suited for research and commercial applications requiring advanced multimodal understanding and high model throughput.

$0.15 / 1M Input Price
$0.60 / 1M Output Price
1M tokens Context Window
16,384 tokens Max Output

Capabilities

Tool Use Function Calling Vision

Architecture

ModalityText + Image → Text
TokenizerLlama4

Performance Indices

Source: Artificial Analysis

18.4 Intelligence Index
15.6 Coding Index
12.3 Agentic Index
19.3 Math Index

Benchmark Scores

Evaluations

GPQA Diamond 67.1%
Graduate-level scientific reasoning
HLE 4.8%
Humanity's Last Exam
MMLU Pro 80.9%
Multi-task language understanding
LiveCodeBench 39.7%
Live coding evaluation
SciCode 33.1%
Scientific computing
MATH 500 88.9%
Mathematical problem-solving
AIME 39%
Competition mathematics
AIME 2025 19.3%
Competition mathematics (2025)
IFBench 43%
Instruction following
LCR 46%
Long-context reasoning
TerminalBench Hard 6.8%
Agentic terminal tasks
τ²-Bench 17.8%
Conversational agent benchmark

Benchmark data from Artificial Analysis and Hugging Face

Model Information

OpenRouter ID meta-llama/llama-4-maverick
Providermeta-llama
Model FamilyLlama 4
Release Date April 5, 2025
Context Length1,048,576 tokens
Max Completion16,384 tokens
Status Active

Pricing

Token Type Cost per 1M tokens Cost per 1K tokens
Input $0.15 $0.000150
Output $0.60 $0.000600

Live Performance

Live endpoint metrics — refreshed every 30 minutes.

97.4%
Avg Uptime
286ms
Best Latency (TTFT)
156 tok/s
Best Throughput
7/7
Active Endpoints
Available via: DeepInfra, Novita, Together, Parasail, Google, SambaNova

Leaderboard Categories