ApX logoApX logo

GLM-5

Active Parameters

744B

Context Length

204.8K

Modality

Multimodal

Architecture

Mixture of Experts (MoE)

License

MIT

Release Date

12 Feb 2026

Knowledge Cutoff

Dec 2025

Technical Specifications

Total Expert Parameters

40.0B

Number of Experts

256

Active Experts

8

Attention Structure

Multi-Head Attention

Hidden Dimension Size

-

Number of Layers

80

Attention Heads

-

Key-Value Heads

-

Activation Function

-

Normalization

RMS Normalization

Position Embedding

Absolute Position Embedding

GLM-5

GLM-5 is a flagship multimodal foundation model developed by Z.ai, designed for complex systems engineering and long-horizon agentic workflows. Utilizing a Mixture-of-Experts (MoE) architecture, the model scales to 744 billion total parameters with approximately 40 billion parameters activated per token. This design facilitates high-capacity reasoning and specialized knowledge retrieval while maintaining the computational efficiency required for large-scale deployment. The model is trained on a massive 28.5 trillion token corpus, emphasizing high-quality code, technical documentation, and reasoning-dense data to support professional-grade software development and autonomous problem-solving.

Technically, GLM-5 introduces several architectural innovations, most notably the integration of DeepSeek Sparse Attention (DSA). This mechanism optimizes the standard attention block by dynamically allocating computational resources, which significantly reduces the memory and compute overhead associated with processing long sequences. Additionally, the model leverages an asynchronous reinforcement learning infrastructure known as 'slime' during post-training. This framework decouples generation from training to improve iteration throughput, allowing the model to learn effectively from complex, multi-step interactions and dynamic environments.

Optimized for long-context stability, GLM-5 supports a context window of up to 204,800 tokens and is capable of generating up to 128,000 tokens in a single output. Its operational capabilities include advanced tool-use, real-time streaming, and structured output across frontend, backend, and data processing tasks. The model is released with open weights under the MIT License, enabling researchers and developers to perform local serving, fine-tuning, and integration into diverse agentic frameworks without vendor lock-in.

About GLM 5

GLM 5 is the fifth generation of General Language Models developed by Z.ai. It represents a significant leap in multimodal foundational capabilities, featuring advanced reasoning and long-horizon agentic capabilities across diverse systems engineering tasks.


Other GLM 5 Models
  • No related models available

Evaluation Benchmarks

Rank

#21

BenchmarkScoreRank

Web Development

WebDev Arena

1447

9

Agentic Coding

LiveBench Agentic

0.55

10

Professional Knowledge

MMLU Pro

0.86

13

0.83

15

0.68

15

0.55

19

0.74

23

0.69

24

Rankings

Overall Rank

#21

Coding Rank

#34

Model Transparency

Total Score

B+

79 / 100

GPU Requirements

Full Calculator

Choose the quantization method for model weights

Context Size: 1,024 tokens

1k
100k
200k

VRAM Required:

Recommended GPUs

GLM-5: Specifications and GPU VRAM Requirements