NVIDIA: Llama 3.1 Nemotron 70B Instruct
NVIDIA's Llama 3.1 Nemotron 70B is a well-rounded model with tool use, function calling, and a large 131K context window, showing solid benchmark performance across reasoning and math. However, at $1.2/1M tokens it faces stiff competition from cheaper alternatives with comparable scores.
Assessment date: March 14, 2026
Our methodology takes into account a range of factors including pricing, functionality, capabilities, benchmark performance, and real-world applicability. Rankings are reviewed and updated regularly as new models are released. Issues with our rankings? Contact us
NVIDIA's Llama 3.1 Nemotron 70B is a language model designed for generating precise and useful responses. Leveraging Llama 3.1 70B architecture and Reinforcement Learning from Human Feedback (RLHF), it excels in automatic alignment benchmarks. This model is tailored for applications requiring high accuracy in helpfulness and response generation, suitable for diverse user queries across multiple domains. Usage of this model is subject to Meta's Acceptable Use Policy.
Capabilities
Architecture
| Modality | Text → Text |
| Tokenizer | Llama3 |
| Instruct Type | llama3 |
| Parameters | 70B |
Performance Indices
Source: Artificial Analysis
Benchmark Scores
Evaluations
Benchmark data from Artificial Analysis and Hugging Face
Model Information
Pricing
| Token Type | Cost per 1M tokens | Cost per 1K tokens |
|---|---|---|
| Input | $1.20 | $0.001200 |
| Output | $1.20 | $0.001200 |
Live Performance
Live endpoint metrics — refreshed every 30 minutes.
Leaderboard Categories
External Resources
Data sourced from OpenRouter API, Artificial Analysis and Hugging Face Open LLM Leaderboard. Scores are editorially curated by our team.
Last updated: March 15, 2026 7:52 pm