inclusionAI/Ling-2.6-1T · Hugging Face

Reddit r/LocalLLaMA / 4/30/2026

📰 NewsTools & Practical UsageIndustry & Market MovesModels & Research

Key Points

  • Hugging Faceで、Lingファミリーの「inclusionAI/Ling-2.6-1T」がオープンソースとして公開され、1兆パラメータ級のフラッグシップモデルとして複雑な実タスクを狙うとしています。
  • 推論効率ではMLAとLinear Attentionのハイブリッド構成により、長文コンテキストでのレイテンシとVRAM使用量を抑えつつ、スループットと計算コストを改善したと説明しています。
  • 「Fast Thinking」として、ポストトレーニングでContextual Process Redundancy Suppression報酬戦略を導入し、冗長なCoT(思考の連鎖)への依存を下げて出力コストを圧縮しながら高い知能を維持するとしています。
  • 推論・エージェント実行面では、コードと日常ワークフローに強く、実行負荷ベンチマーク(AIME26、SWE-bench Verified、BFCL-V4、TAU2-Bench、IFBenchなど)でオープンソースSOTAを達成し、Claude CodeやOpenClaw等の主要エージェント系フレームワークと統合しやすいとしています。
inclusionAI/Ling-2.6-1T · Hugging Face

Ling-2.6-1T: A Trillion-Parameter Comprehensive Flagship Model for Complex Tasks

Today, we are thrilled to open-source Ling–2.6–1T from the Ling family.

Tailored for real–world, complex scenarios, this trillion–parameter model introduces targeted optimizations across inference efficiency, token overhead, and agentic capabilities, making it highly effective for coding and daily workflows.

Key upgrades in Ling–2.6–1T include:

  • High Inference Efficiency: By adopting a hybrid architecture combining MLA and Linear Attention, we dramatically reduce latency and VRAM footprint for long contexts. It delivers superior throughput and lower per–token computational costs without sacrificing expressivity, ensuring real–time responsiveness for complex reasoning and tool calling.
  • Lower Token Overhead via "Fast Thinking": We introduce a Contextual Process Redundancy Suppression reward strategy during post–training. This reduces reliance on verbose chains–of–thought (CoT), utilizing a "fast thinking" mechanism to reach answers directly and compress output costs while maintaining top–tier intelligence.
  • Reliable Multi–Step Execution: With enhanced reasoning, agentic coding, and instruction following, Ling–2.6–1T achieves open–source SOTA on execution–heavy benchmarks, including AIME26, SWE–bench Verified, BFCL–V4, TAU2–Bench, and IFBench.
  • Production–Ready for Agent Workflows: Designed for end–to–end engineering—from code generation to bug fixing—Ling–2.6–1T integrates seamlessly with mainstream agent frameworks like Claude Code, OpenClaw, OpenCode, and CodeBuddy, effortlessly handling multi–tool, multi–step constraints in enterprise environments.
submitted by /u/pmttyji
[link] [comments]