AI Disruption

AI Disruption

Zhipu’s New Model Also Uses DeepSeek’s MLA, Runs on Apple M5

Zhipu IPO follow-up: free 30B MoE GLM-4.7-Flash, 3B active, MLA power, M5 local 43 tok/s

Meng Li's avatar
Meng Li
Jan 20, 2026
∙ Paid

“AI Disruption” Publication 8600 Subscriptions 20% Discount Offer Link.


Zhipu AI Launches New Achievement After IPO.

Zhipu AI has released a new open-source lightweight large language model, GLM-4.7-Flash, which directly replaces the previous generation GLM-4.5-Flash, with API access freely available.

This is a Mixture of Experts (MoE) architecture model with 30B total parameters and only 3B activated parameters. The official positioning for it is “local programming and intelligent agent assistant.”

In the SWE-bench Verified code fixing test, GLM-4.7-Flash achieved a score of 59.2, and also significantly outperformed models of similar scale, such as Qwen3-30B and GPT-OSS-20B in evaluations including “humanity’s last exam.”

User's avatar

Continue reading this post for free, courtesy of Meng Li.

Or purchase a paid subscription.
© 2026 Meng Li · Privacy ∙ Terms ∙ Collection notice
Start your SubstackGet the app
Substack is the home for great culture