Mistral: Mistral Small 3

Mistral: Mistral Small 3

mistralai · Released Jan 30, 2025
38
Our Score

Mistral Small 3 is a 24B-parameter language model optimized for low-latency performance across common AI tasks. Released under the Apache 2.0 license, it features both pre-trained and instruction-tuned versions designed for efficient local deployment. The model achieves 81% accuracy on the MMLU benchmark and performs competitively with larger models like Llama 3.3 70B and Qwen 32B, while operating at three times the speed on equivalent hardware. Read the blog post about the model here.

$0.05 / 1M Input Price
$0.08 / 1M Output Price
32,768 tokens Context Window
16,384 tokens Max Output
24B Parameters

Capabilities

Tool Use Function Calling

Architecture

ModalityText → Text
TokenizerMistral
Parameters24B

Performance Indices

Source: Artificial Analysis

12.7 Intelligence Index
19.6 Agentic Index
4.3 Math Index

Benchmark Scores

Evaluations

GPQA Diamond 46.2%
Graduate-level scientific reasoning
HLE 4.1%
Humanity's Last Exam
MMLU Pro 65.2%
Multi-task language understanding
LiveCodeBench 25.2%
Live coding evaluation
SciCode 23.6%
Scientific computing
MATH 500 71.5%
Mathematical problem-solving
AIME 8%
Competition mathematics
AIME 2025 4.3%
Competition mathematics (2025)
IFBench 26.4%
Instruction following
τ²-Bench 19.6%
Conversational agent benchmark

Benchmark data from Artificial Analysis and Hugging Face

Model Information

OpenRouter ID mistralai/mistral-small-24b-instruct-2501
Providermistralai
Model FamilyMistral
Release Date January 30, 2025
Context Length32,768 tokens
Max Completion16,384 tokens
Status Active

Pricing

Token Type Cost per 1M tokens Cost per 1K tokens
Input $0.05 $0.000050
Output $0.08 $0.000080

Live Performance

Live endpoint metrics — refreshed every 30 minutes.

100%
Avg Uptime
658ms
Best Latency (TTFT)
49 tok/s
Best Throughput
1/2
Active Endpoints
Available via: DeepInfra, Together

Leaderboard Categories