Meta: Llama 3.1 70B Instruct
Llama 3.1 70B Instruct from Meta offers a solid balance of capability and cost, with a 131K context window, tool use, and reasonable benchmark scores across reasoning and coding; it is a dependable open-weight option for content and tool-use workflows at a competitive price.
Assessment date: March 12, 2026
Our methodology takes into account a range of factors including pricing, functionality, capabilities, benchmark performance, and real-world applicability. Rankings are reviewed and updated regularly as new models are released. Issues with our rankings? Contact us
Meta's latest class of model (Llama 3.1) launched with a variety of sizes & flavors. This 70B instruct-tuned version is optimized for high quality dialogue usecases. It has demonstrated strong performance compared to leading closed-source models in human evaluations. Usage of this model is subject to Meta's Acceptable Use Policy.
Capabilities
Architecture
| Modality | Text → Text |
| Tokenizer | Llama3 |
| Instruct Type | llama3 |
| Parameters | 70B |
Performance Indices
Source: Artificial Analysis
Benchmark Scores
Evaluations
Benchmark data from Artificial Analysis and Hugging Face
Model Information
Pricing
| Token Type | Cost per 1M tokens | Cost per 1K tokens |
|---|---|---|
| Input | $0.40 | $0.000400 |
| Output | $0.40 | $0.000400 |
Live Performance
Live endpoint metrics — refreshed every 30 minutes.
Leaderboard Categories
External Resources
Data sourced from OpenRouter API, Artificial Analysis and Hugging Face Open LLM Leaderboard. Scores are editorially curated by our team.
Last updated: March 13, 2026 7:52 pm