inclusionAI: Ling-2.6-flash

inclusionAI: Ling-2.6-flash

inclusionai · Released Apr 21, 2026 New
Intelligence
— Awaiting review
Speed #24 / 246
206.6 tokens / sec
Input #165 / 538
$0.080 per 1M tokens
Output #177 / 538
$0.240 per 1M tokens
Context #87 / 538
262,144 tokens

Analysis Summary

At $0.080 input and $0.240 output per 1M tokens, it is among the most expensive on the market. It offers a generous context window for extended reasoning and code review and supports tool use and function calling.

Rankings consider pricing, capabilities, benchmarks, and real-world applicability and are refreshed as new models launch. Feedback?

Ling-2.6-flash is an instant (instruct) model from inclusionAI with 104B total parameters and 7.4B active parameters, designed for real-world agents that require fast responses, strong execution, and high token efficiency..

Capabilities

Tool Use Function Calling

Performance Indices

Source: Artificial Analysis

26.2 Intelligence Index
23.2 Coding Index
53.6 Agentic Index

This model was released recently. Independent benchmark evaluations are typically completed within days of release — these figures are preliminary and are likely to be updated as testing is finalised.

Benchmark Scores

Intelligence

GPQA Diamond 59.3% Graduate-level scientific reasoning
HLE 6.2% Humanity's Last Exam
SciCode 27.1% Scientific computing

Technical

TerminalBench Hard 21.2% Agentic terminal tasks
τ²-Bench 86% Conversational agent benchmark

Content

IFBench 57.4% Instruction following
LCR 25% Long-context reasoning

Benchmark data from Artificial Analysis and Hugging Face

How does inclusionAI: Ling-2.6-flash stack up?

Compare side-by-side with other similar models.

Compare Models

Model Information

OpenRouter ID inclusionai/ling-2.6-flash
Providerinclusionai
Release Date April 21, 2026
Context Length262,144 tokens
Max Completion32,768 tokens
Status Active

Pricing

Token Type Cost per 1M tokens Cost per 1K tokens
Input $0.08 $0.000080
Output $0.24 $0.000240