Skip to main content

Zhipu AI Unveils GLM-4.5: Open-Source AI with Integrated Reasoning

Zhipu AI Unveils GLM-4.5: A New Benchmark in Open-Source AI

Beijing, July 28, 2025 — Zhipu AI has officially launched GLM-4.5, its next-generation flagship model designed specifically for agent applications. The model is now open-sourced on Hugging Face and ModelScope, with weights licensed under the MIT License, making it accessible to developers worldwide.

Open-Source SOTA Performance

GLM-4.5 sets a new standard for state-of-the-art (SOTA) performance in reasoning, coding, and agent capabilities. In real-world evaluations, it outperformed all other domestic models. Comprehensive testing across 12 benchmark datasets—including MMLU Pro, AIME24, and MATH500—placed GLM-4.5 third globally, first domestically, and first among open-source models.

Image

Native Integration of Core Capabilities

The model is the first to achieve native integration of reasoning, coding, and agent functionalities. This breakthrough aligns with Zhipu AI's vision for AGI (Artificial General Intelligence), ensuring that new capabilities do not compromise existing ones.

High-Parameter-Efficiency MoE Architecture

GLM-4.5 employs a Mixture of Experts (MoE) architecture, available in two versions:

  • GLM-4.5: 355 billion total parameters (32 billion activated).
  • GLM-4.5-Air: 106 billion total parameters (12 billion activated).

Despite having fewer parameters than competitors like DeepSeek-R1 and Kimi-K2, GLM-4.5 delivers superior performance on multiple benchmarks.

Image

Two-Mode Design and Training Process

The model supports:

  1. Thinking Mode: For complex reasoning and tool usage.
  2. Non-Thinking Mode: For instant responses.

The training process involved:

  • Pre-training on 15 trillion tokens of general data.
  • Targeted training on 8 trillion tokens for code, reasoning, and agents.
  • Reinforcement learning to enhance capabilities.

Image

Cost-Effective API Pricing

GLM-4.5 offers industry-leading pricing:

  • Input: 0.8 yuan per million tokens.
  • Output: 2 yuan per million tokens. The high-speed version generates up to 100 tokens per second, supporting low-latency applications.

Image

Full-Stack Development & Ecosystem Compatibility

The model excels in full-stack tasks, from front-end website writing to back-end database management. It is optimized for frameworks like Claude Code, Cline, and Roo Code, offering one-click compatibility with Claude Code.

Multi-Platform Access & Transparency

Users can access GLM-4.5 via:

  • The BigModel.cn open platform.
  • Free features on Zhipu Qingyan (chatglm.cn) and z.ai. To ensure transparency, Zhipu AI has published 52 questions and Agent trajectories for verification.

The release of GLM-4.5 is poised to drive large-scale AGI adoption across industries.

Key Points:

  1. GLM-4.5 is the first open-source model with native reasoning, coding, and agent integration.
  2. Achieves SOTA performance with fewer parameters than competitors.
  3. Offers cost-efficient API pricing at 0.8 yuan per million input tokens.
  4. Supports full-stack development and major code agent frameworks.

Enjoyed this article?

Subscribe to our newsletter for the latest AI news, product reviews, and project recommendations delivered to your inbox weekly.

Weekly digestFree foreverUnsubscribe anytime

Related Articles

Musk's xAI Launches Grok Imagine 1.0, Bringing HD AI Video to the Masses
News

Musk's xAI Launches Grok Imagine 1.0, Bringing HD AI Video to the Masses

Elon Musk's AI venture xAI has officially released Grok Imagine 1.0, marking a significant leap in AI-generated video quality. The new version produces crisp 720p videos up to 10 seconds long with improved audio - a notable upgrade from previous offerings. During testing, users created over 1.2 billion videos, showing strong demand for accessible video generation tools.

February 2, 2026
Artificial IntelligenceVideo GenerationElon Musk
News

AI Pioneer David Silver Strikes Out Alone With Bold New Vision

David Silver, the mastermind behind DeepMind's groundbreaking AlphaGo, has left Google to pursue his own AI venture. His new startup, Ineffable Intelligence, challenges the current obsession with large language models by betting on reinforcement learning - arguing that true intelligence comes from experience, not just data. This move signals a growing divide in AI research approaches.

February 2, 2026
Artificial IntelligenceReinforcement LearningTech Startups
Ant Group Bets Big on AI with New Incentive Program
News

Ant Group Bets Big on AI with New Incentive Program

Ant Group is doubling down on artificial intelligence with its newly launched 'AI Credit' program. The initiative rewards teams making significant AI advancements, offering additional incentives convertible to long-term economic benefits. While celebrating milestones like their trillion-parameter Bailin models and popular AI assistant Lingguang, CEO Han Xinyi maintains Ant still trails industry leaders. The company sees payments, finance, and healthcare as key growth areas for its AI transformation.

February 2, 2026
Artificial IntelligenceFintech InnovationCorporate Strategy
News

Apple's AI Brain Drain: Top Siri Exec and Researchers Jump Ship

Apple's artificial intelligence team is bleeding talent as key researchers and executives depart for rivals like Meta and Google DeepMind. The exodus includes Siri's former senior director Stuart Bowers and four prominent AI scientists, raising questions about Apple's ability to retain top tech minds in the fiercely competitive AI landscape.

February 2, 2026
AppleArtificial IntelligenceTech Talent Wars
Tsinghua AI Whiz Joins Tencent to Supercharge Multimodal Learning
News

Tsinghua AI Whiz Joins Tencent to Supercharge Multimodal Learning

Tencent's AI ambitions get a major boost as Peng Tianyu, a rising star in machine learning from Tsinghua University, joins their Tongyi team. The 31-year-old prodigy brings expertise in reinforcement learning and multimodal systems, fresh from his stint at Sea AI Lab in Singapore. This marks another strategic hire for Tencent following their recent acquisition of an OpenAI researcher.

January 30, 2026
TencentAI ResearchMachine Learning
News

SenseTime's New AI Model Thinks Like a Detective

SenseTime has unveiled SenseNova-MARS, an open-source AI model that combines visual reasoning with text-image search capabilities. Outperforming GPT-5.2 on multiple benchmarks, this innovative technology mimics human-like investigation skills - zooming in on tiny details, connecting information dots, and solving complex problems autonomously. The company has made both the 8B and 32B versions publicly available for developers worldwide.

January 30, 2026
AI InnovationComputer VisionMachine Learning