Back to stories
Models

Zhipu AI Releases GLM-5: A 744B Parameter Model Under MIT License

Michael Ouroumis2 min read
Zhipu AI Releases GLM-5: A 744B Parameter Model Under MIT License

Chinese AI lab Zhipu AI has released GLM-5, a 744-billion parameter mixture-of-experts (MoE) model, under the MIT license. The model uses 44 billion active parameters per forward pass, features a 200,000-token context window, and scores an impressive 77.8% on SWE-bench Verified.

Model Specifications

SpecValue
Total parameters744B
Active parameters44B (MoE)
Context window200K tokens
SWE-bench Verified77.8%
Training hardwareHuawei Ascend
LicenseMIT

Why This Matters

Open-Source Under MIT

The MIT license is the most permissive available — anyone can use, modify, and commercialize the model without restrictions. This is a stronger commitment to open-source than Meta's Llama license, which includes usage-based restrictions for large companies.

Trained on Huawei Ascend

GLM-5 was trained entirely on Huawei's Ascend AI chips rather than NVIDIA GPUs. This is significant because US export controls have restricted China's access to advanced NVIDIA hardware. GLM-5's strong benchmark performance demonstrates that the Chinese AI ecosystem can produce competitive models despite these restrictions.

SWE-bench Performance

The 77.8% score on SWE-bench Verified — a benchmark that tests a model's ability to solve real-world software engineering tasks from GitHub issues — places GLM-5 among the top-performing models on this widely-watched benchmark, alongside tools like Moonshot's Kimi Code which scored 62% on the same test.

The Chinese Open-Source Wave

GLM-5 is the latest in a series of strong open-source releases from Chinese AI labs:

This trend is creating a two-track open-source ecosystem, with both Western labs (Meta, Mistral) and Chinese labs (Zhipu, DeepSeek, Alibaba) producing frontier-quality open models. Alibaba's Qwen3.5, for example, offers 201-language support with an agent-first architecture.

Availability

The model weights, tokenizer, and training documentation are available on Hugging Face and ModelScope. Zhipu has also published a technical report detailing the architecture, training process, and evaluation methodology.

Community members have already begun creating quantized versions for consumer hardware, though the full model requires significant compute resources to run.

Learn AI for Free — FreeAcademy.ai

Take "AI Essentials: Understanding AI in 2026" — a free course with certificate to master the skills behind this story.

More in Models

Moonshot Kimi K2.6 lands open-source, scales to 300 sub-agents and 4,000 coordinated steps
Models

Moonshot Kimi K2.6 lands open-source, scales to 300 sub-agents and 4,000 coordinated steps

Moonshot AI shipped Kimi K2.6 as a generally available open-source model on April 20, posting 58.6 on SWE-Bench Pro — ahead of GPT-5.4 and Claude Opus 4.6 — while scaling agent swarms to 300 sub-agents and 4,000 coordinated steps.

9 hours ago3 min read
OpenAI's 'Spud' Caught Live in API Testing, Polymarket Jumps to 81% for April 23 Launch
Models

OpenAI's 'Spud' Caught Live in API Testing, Polymarket Jumps to 81% for April 23 Launch

API monitors detected OpenAI's next frontier model — codenamed Spud — running in production-scale testing on April 19, sending Polymarket traders to an 81% implied probability of a public launch on April 23.

1 day ago2 min read
OpenAI Launches GPT-Rosalind, Its First Domain-Specific Model Built for Life Sciences
Models

OpenAI Launches GPT-Rosalind, Its First Domain-Specific Model Built for Life Sciences

OpenAI debuts GPT-Rosalind, a specialized AI model for biology, drug discovery, and genomics, with launch partners including Amgen, Moderna, and Los Alamos National Laboratory.

4 days ago2 min read