inclusionAI: Ling-2.6-flash (free)

inclusionAI: Ling-2.6-flash (free)

inclusionai · Released Apr 21, 2026 Efficient New
Intelligence #230 / 525
35.6 Our Score
Speed #24 / 244
210.4 tokens / sec
Input
Not priced
Output
Not priced
Context #80 / 525
262,144 tokens

Analysis Summary

inclusionAI: Ling-2.6-flash (free) sits in the Efficient tier on our leaderboard, ranked #230 of 525 published models on overall intelligence. At $0.000 input and $0.000 output per 1M tokens, it is among the most expensive on the market. It offers a generous context window for extended reasoning and code review and supports tool use and function calling.

Editorial notes

InclusionAI's Ling-2.6-Flash is a free model with tool use and a 262K context window, offering decent agentic capability for its tier, though its reasoning and coding scores are modest and best suited to lighter tasks.

Assessed April 23, 2026

Rankings consider pricing, capabilities, benchmarks, and real-world applicability and are refreshed as new models launch. Feedback?

Performance Profile

Intelligence4.4Technical5.2Value0Content4.5
Intelligence 4.4/10
Technical 5.2/10
Content 4.5/10
Value 0/10

Ling-2.6-flash is an instant (instruct) model from inclusionAI with 104B total parameters and 7.4B active parameters, designed for real-world agents that require fast responses, strong execution, and high token efficiency..

Capabilities

Tool Use Function Calling

Performance Indices

Source: Artificial Analysis

26.2 Intelligence Index
23.2 Coding Index
53.6 Agentic Index

This model was released recently. Independent benchmark evaluations are typically completed within days of release — these figures are preliminary and are likely to be updated as testing is finalised.

Benchmark Scores

Intelligence

GPQA Diamond 59.3% Graduate-level scientific reasoning
HLE 6.2% Humanity's Last Exam
SciCode 27.1% Scientific computing

Technical

TerminalBench Hard 21.2% Agentic terminal tasks
τ²-Bench 86% Conversational agent benchmark

Content

IFBench 57.4% Instruction following
LCR 25% Long-context reasoning

Benchmark data from Artificial Analysis and Hugging Face

How does inclusionAI: Ling-2.6-flash (free) stack up?

Compare side-by-side with other efficient models.

Compare Models

Model Information

OpenRouter ID inclusionai/ling-2.6-flash:free
Providerinclusionai
Release Date April 21, 2026
Context Length262,144 tokens
Max Completion32,768 tokens
Status Active

Live Performance

Live endpoint metrics — refreshed every 30 minutes.

100%
Avg Uptime
1,146ms
Best Latency (TTFT)
55 tok/s
Best Throughput
1/1
Active Endpoints
Available via: Novita