ApX 标志ApX 标志

趋近智

GLM-4-9B-Chat

参数

9B

上下文长度

128K

模态

Text

架构

Dense

许可证

MIT License

发布日期

30 Jun 2024

训练数据截止日期

Dec 2023

技术规格

注意力结构

Multi-Head Attention

隐藏维度大小

4096

层数

40

注意力头

32

键值头

2

激活函数

SwigLU

归一化

RMS Normalization

位置嵌入

Absolute Position Embedding

GLM-4-9B-Chat

The GLM-4-9B-Chat model is a conversational large language model developed by the Knowledge Engineering Group (KEG) at Tsinghua University in collaboration with Z.ai. As a core component of the fourth-generation General Language Model (GLM) series, this variant is specifically optimized for human-preference alignment and complex multi-turn dialogue. The model is trained on a massive corpus of 10 trillion tokens and supports multilingual communication across 26 languages, making it a highly versatile tool for global conversational applications.

Architecturally, GLM-4-9B-Chat is built on a dense transformer framework utilizing 40 layers with a hidden dimension of 4096. A significant technical innovation in this variant is the implementation of Grouped Query Attention (GQA), which employs two key-value heads to optimize memory bandwidth and inference throughput without sacrificing modeling quality. The architecture further incorporates Rotary Position Embeddings (RoPE) for improved length extrapolation and utilizes SwiGLU activation functions in its feed-forward networks, replacing traditional ReLU to enhance the model's non-linear representative capacity. Normalized using RMSNorm, the model maintains stable training dynamics across its parameter space.

GLM-4-9B-Chat is engineered to handle extended context windows up to 128,000 tokens, enabling it to maintain coherence over long documents and extensive conversational histories. Beyond standard text generation, the model integrates sophisticated tool-use capabilities, including autonomous web browsing, Python code execution, and custom function calling. These features allow the model to interact with external environments to solve mathematical problems and perform real-time information retrieval, making it suitable for deployment in advanced AI assistants and automated agentic systems.

关于 GLM Family

General Language Models from Z.ai


其他 GLM Family 模型

评估基准

没有可用的 GLM-4-9B-Chat 评估基准。

排名

排名

-

编程排名

-

GPU 要求

完整计算器

选择模型权重的量化方法

上下文大小:1024 个令牌

1k
63k
125k

所需显存:

推荐 GPU