Google: Gemma 3n 2B (free)

Google: Gemma 3n 2B (free)

google · Released Jul 9, 2025
18
Our Score

Gemma 3n E2B IT is a multimodal, instruction-tuned model developed by Google DeepMind, designed to operate efficiently at an effective parameter size of 2B while leveraging a 6B architecture. Based on the MatFormer architecture, it supports nested submodels and modular composition via the Mix-and-Match framework. Gemma 3n models are optimized for low-resource deployment, offering 32K context length and strong multilingual and reasoning performance across common benchmarks. This variant is trained on a diverse corpus including code, math, web, and multimodal data.

8,192 tokens Context Window
2,048 tokens Max Output
2B Parameters

Architecture

ModalityText → Text
TokenizerOther
Parameters2B

Performance Indices

Source: Artificial Analysis

4.8 Intelligence Index
2.2 Coding Index
0.8 Agentic Index
10.3 Math Index

Benchmark Scores

Evaluations

GPQA Diamond 22.9%
Graduate-level scientific reasoning
HLE 4%
Humanity's Last Exam
MMLU Pro 37.8%
Multi-task language understanding
LiveCodeBench 9.5%
Live coding evaluation
SciCode 5.2%
Scientific computing
MATH 500 69.1%
Mathematical problem-solving
AIME 9%
Competition mathematics
AIME 2025 10.3%
Competition mathematics (2025)
IFBench 22%
Instruction following
TerminalBench Hard 0.8%
Agentic terminal tasks

Benchmark data from Artificial Analysis and Hugging Face

Model Information

OpenRouter ID google/gemma-3n-e2b-it:free
Providergoogle
Release Date July 9, 2025
Context Length8,192 tokens
Max Completion2,048 tokens
Status Active

Live Performance

Live endpoint metrics — refreshed every 30 minutes.

100%
Avg Uptime
389ms
Best Latency (TTFT)
52 tok/s
Best Throughput
1/1
Active Endpoints
Available via: Google AI Studio