ApX 标志

趋近智

MiniMax M2

活跃参数

229B

上下文长度

128K

模态

Text

架构

Mixture of Experts (MoE)

许可证

MIT

发布日期

7 Nov 2025

知识截止

-

技术规格

专家参数总数

10.0B

专家数量

-

活跃专家

2

注意力结构

Multi-Head Attention

隐藏维度大小

-

层数

-

注意力头

-

键值头

-

激活函数

-

归一化

-

位置嵌入

Absolute Position Embedding

系统要求

不同量化方法和上下文大小的显存要求

MiniMax M2

MiniMax M2 is a Mixture of Experts (MoE) model developed by MiniMax, engineered for high performance in coding and agentic tasks. The model is designed to deliver advanced capabilities while optimizing cost and inference speed, making it suitable for practical deployment. It supports end-to-end developer workflows, including multi-file edits, code-run-fix loops, and long-horizon toolchains.

Architecturally, MiniMax M2 employs a sparse MoE transformer design, comprising a total of 230 billion parameters. A key innovation lies in its efficient activation strategy, where only 10 billion parameters are actively utilized during inference for each token. This selective activation mechanism reduces computational demands significantly while maintaining a broad capacity for knowledge and reasoning. The model's architecture is also characterized as a "full attention model," implying the use of a Multi-Head Attention (MHA) mechanism. Furthermore, MiniMax M2 supports multimodal inputs, encompassing text, audio, images, and video, extending its applicability across diverse data types.

Purpose-built for AI agent workflows and coding tasks, MiniMax M2 provides native support for integrating external tools such as shell environments, web browsers, and Python interpreters. This enables the model to facilitate complex, multi-step processes and robust tool-calling sequences. The model's efficiency allows for flexible deployment across various inference frameworks. Its design supports fast feedback loops, a critical attribute for environments like integrated development environments (IDEs) and continuous integration (CI) pipelines. An important operational aspect is the model's ability to maintain reasoning traces between turns, which is integral for consistent agent performance and improved auditability of its decision-making processes.

关于 MiniMax M2

MiniMax's efficient MoE models built for coding and agentic workflows.


其他 MiniMax M2 模型
  • 没有相关模型

评估基准

排名适用于本地LLM。

没有可用的 MiniMax M2 评估基准。

排名

排名

-

编程排名

-

GPU 要求

完整计算器

选择模型权重的量化方法

上下文大小:1024 个令牌

1k
63k
125k

所需显存:

推荐 GPU