趋近智
活跃参数
358B
上下文长度
200K
模态
Text
架构
Mixture of Experts (MoE)
许可证
MIT
发布日期
8 Jan 2026
训练数据截止日期
Sep 2024
专家参数总数
32.0B
专家数量
-
活跃专家
-
注意力结构
Multi-Head Attention
隐藏维度大小
-
层数
-
注意力头
-
键值头
-
激活函数
-
归一化
RMS Normalization
位置嵌入
Absolute Position Embedding
GLM-4.7 is a large-scale Mixture of Experts (MoE) model developed by Z.ai, specifically architected to support advanced agentic coding, complex reasoning, and multi-step tool orchestration. Building upon the GLM-4 series, the model integrates a sophisticated reasoning system that prioritizes logical consistency and task completion across extended interactions. It is designed to function as a primary engine for coding agents and terminal-based automation, featuring optimizations for multi-language programming and autonomous execution within complex software environments.
The model's technical foundation includes a triple-tier thinking architecture designed to maintain reasoning coherence. Interleaved Thinking allows the model to perform internal reasoning steps before every response and tool invocation, ensuring that generated instructions align with logical constraints. Preserved Thinking facilitates the retention of these reasoning blocks across multi-turn conversations, preventing the context decay typically seen in long-horizon tasks. Additionally, Turn-level Thinking provides a granular control mechanism, allowing developers to adjust reasoning depth based on the specific requirements of each interaction to manage computational overhead and latency effectively.
Beyond programming, GLM-4.7 features a refined approach to frontend and user interface development, often referred to as vibe coding. This capability focuses on generating aesthetically consistent and structurally sound UI code, including modern web pages and professional presentation layouts. The model's architecture also emphasizes robust tool integration, enabling it to navigate terminal environments, execute shell commands, and interact with external APIs while maintaining a high degree of stability and instruction adherence in diverse automation scenarios.
GLM-4 is a series of bilingual (English and Chinese) language models developed by Zhipu AI. The models feature extended context windows, superior coding performance, advanced reasoning capabilities, and strong agent functionalities. GLM-4.6 offers improvements in tool use and search-based agents.
排名
#18
| 基准 | 分数 | 排名 |
|---|---|---|
Data Analysis LiveBench Data Analysis | 0.74 | ⭐ 4 |
Professional Knowledge MMLU Pro | 0.84 | 4 |
Graduate-Level QA GPQA | 0.86 | 8 |
Web Development WebDev Arena | 1441 | 11 |
Agentic Coding LiveBench Agentic | 0.42 | 16 |
Coding LiveBench Coding | 0.73 | 18 |
Mathematics LiveBench Mathematics | 0.76 | 20 |
Reasoning LiveBench Reasoning | 0.60 | 22 |