跳转到主要内容

Ant Group's BaiLing Team Open Sources Efficient AI Model

Ant Group's BaiLing Team Releases Revolutionary AI Model

Amid fierce competition in AI development, Ant Group's BaiLing large model team has open-sourced Ring-flash-linear-2.0-128K, a groundbreaking model designed specifically for ultra-long text programming applications. This release marks a significant advancement in efficient AI inference and long-context processing.

Image

Hybrid Architecture Delivers Unprecedented Efficiency

The model features an innovative hybrid linear + standard attention mechanism combined with a sparse MoE (Mixture of Experts) architecture. With total parameters scaled at 104B but only 6.1B activated during operation (4.8B excluding embeddings), the system achieves:

  • Near-linear time complexity
  • Constant space complexity
  • Generation speeds exceeding 200 tokens/second at 128K context on H20 hardware
  • Three times faster daily use speeds compared to traditional models

The architecture is particularly optimized for resource-limited scenarios while maintaining performance comparable to 40B dense models.

Enhanced Training Yields Superior Reasoning Capabilities

Building upon the Ling-flash-base-2.0 foundation, the model underwent:

  • Additional training on 1T tokens of high-quality data
  • Stable supervised fine-tuning (SFT)
  • Multi-stage reinforcement learning (RL)

The training process overcame traditional instability issues in MoE long-chain reasoning through Ant's proprietary "Icepop" algorithm. Benchmark results demonstrate exceptional capabilities:

  • 86.98 score in AIME2025 math competition
  • 90.23 Elo rating in CodeForces programming tests
  • Outperforms 40B dense models like Qwen3-32B in logical reasoning and creative writing tasks

Image

Long Context Handling Redefines Programming Efficiency

The model natively supports 128K context windows, expandable to 512K using YaRN extrapolation technology. Performance highlights include:

  • Prefill phase throughput nearly 5× higher than Qwen3-32B
  • Decoding phase achieving 10× acceleration
  • Maintains high accuracy even in 32K+ context programming tasks without "model leakage" issues The system proves particularly effective for:

  • Front-end development
  • Structured code generation
  • Agent simulation scenarios

    Open Source Availability Accelerates Adoption

    The BaiLing team has made the model available on:

    /div>">Hugging Face ">ModelScope ">">">">">">">">">">">Support includes BF16/FP8 formats and easy integration with popular frameworks like Transformers, SGLang, and vLLM."";"""Technical documentation is available on arXiv (https://arxiv.org/abs/2510.19338).""",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,""",,,,"",,"",,"",,"",,"",,"",,"",,"",,"",,"".'''''''''''''''',,,,,,,,,,,,''''',,,,,,,,''''',,,,,,,,''''',,,,,,,,''''',,,,,,,,''''',,,,,,,,''''',,,,,,,,''''',,,,,,,,''''',,,,,,,,''''',,,,,,,,''''',,,,,,,,''''',,,,,,,,''''','', '''Key Points:'''''- Combines hybrid linear attention with MoE architecture'- Achieves SOTA performance with only 6.1B activated parameters'- Native 128K context support expandable to 512K'- Sevenfold efficiency improvement over previous versions'- Available now on Hugging Face and ModelScope

喜欢这篇文章?

订阅我们的 Newsletter,获取最新 AI 资讯、产品评测和项目推荐,每周精选直达邮箱。

每周精选完全免费随时退订

相关文章

News

Mac Mini的隐藏实力:工程师如何解锁苹果M4芯片的AI训练能力

在一项令人惊讶的突破中,工程师们破解了苹果神经引擎的能力,揭示Mac Mini不仅能运行应用程序。借助Claude AI的帮助逆向工程M4芯片后,研究人员发现这些紧凑设备能高效训练AI模型——这挑战了对昂贵GPU配置的需求。研究显示其能效比专业级硬件高出80倍,可能实现AI开发的民主化。

March 9, 2026
Apple SiliconAI HardwareMachine Learning
阿里巴巴Qwen3.5 AI模型即将发布,具备视觉语言能力
News

阿里巴巴Qwen3.5 AI模型即将发布,具备视觉语言能力

阿里巴巴新一代AI模型Qwen3.5似乎已准备就绪,代码出现在HuggingFace仓库中。据报道,该模型采用混合注意力机制,并可能作为原生视觉语言模型(VLM)首次亮相。开发者发现了紧凑型2B密集模型和更强大的35B-A3B MoE变体的相关引用。若当前传闻属实,中国春节庆祝活动可能与AI社区这一重要的开源发布同时进行。

February 9, 2026
AIMachine LearningAlibaba
免费AI模型Pony Alpha引发行业热议:它真的是伪装版的GLM-5吗?
News

免费AI模型Pony Alpha引发行业热议:它真的是伪装版的GLM-5吗?

一款名为Pony Alpha的神秘AI模型突然出现在OpenRouter平台上,提供完全免费的强大功能。早期测试显示其在编码和推理任务上可媲美Claude Opus等顶级模型。虽然其来源尚不明确,但许多人怀疑这可能是智谱AI即将推出的GLM-5模型在试水。它的突然出现引发了诸多疑问:为何要免费提供如此先进的技术?开发者又能从用户互动中获得什么?

February 9, 2026
AI ModelsOpenRouterZhipu AI
News

AI的学习鸿沟:为何机器无法像人类一样从失败中成长

一位前OpenAI研究员揭示了当今AI系统的关键缺陷:它们无法从错误中学习。曾协助开发OpenAI关键模型的Jerry Tworek解释道,这种适应能力的缺失威胁着实现真正人工通用智能的进程。与通过试错进化的人类不同,当前AI在面临陌生挑战时会碰壁——这迫使专家们重新思考基础架构。

February 3, 2026
Artificial IntelligenceMachine LearningAGI
小米AI模型开启付费模式:你需要了解的关键信息
News

小米AI模型开启付费模式:你需要了解的关键信息

小米为其MiMo-V2-Flash AI模型推出付费选项的同时仍向用户提供免费额度。这家科技巨头公布了国内外市场的定价细节,该模型拥有惊人的3090亿参数。中国用户需完成实名认证,而海外用户则可通过Apple Pay和信用卡享受更简便的支付方式。

January 21, 2026
XiaomiAI ModelsTech News
AI攻克埃尔德什最难题:GPT5.2的突破令数学家震惊
News

AI攻克埃尔德什最难题:GPT5.2的突破令数学家震惊

GPT5.2在两周内解决了保罗·埃尔德什留下的11个传奇性未解数学难题,并通过形式化证明工具验证。这一突破引起了陶哲轩等顶尖数学家的关注,哈佛大学的诺姆·埃尔基斯更是在AI生成的解决方案基础上继续研究。这标志着人工智能不仅辅助人类研究者,更在纯数学前沿实现自主发现的转折点。

January 15, 2026
Artificial IntelligenceMathematicsGPT5