ApX 标志ApX 标志

趋近智

Sarvam-30B

活跃参数

32B

上下文长度

128K

模态

Text

架构

Mixture of Experts (MoE)

许可证

Apache 2.0

发布日期

6 Mar 2026

训练数据截止日期

-

技术规格

专家参数总数

2.4B

专家数量

128

活跃专家

6

注意力结构

Grouped-Query Attention

隐藏维度大小

4096

层数

19

注意力头

-

键值头

4

激活函数

SwigLU

归一化

RMS Normalization

位置嵌入

ROPE

Sarvam-30B

Sarvam-30B is an advanced Mixture-of-Experts (MoE) model with 32B total parameters and 2.4B active parameters, designed for practical deployment in resource-constrained environments. Released March 6, 2026 under Apache 2.0 license. Uses 19 layers with 128 experts, top-6 routing, grouped KV attention (4 heads), and extremely high rope_theta (8e6) for long-context stability. Delivers state-of-the-art performance across 22 Indian languages with strong reasoning, reliable coding ability, and best-in-class conversational quality. Optimized for multilingual voice calls with tool calling capabilities, throughput, and memory efficiency.

关于 Sarvam

Sarvam AI's sovereign foundation models built for India's languages, culture, and context. Released in March 2026, these advanced Mixture-of-Experts (MoE) models offer state-of-the-art performance across 22 Indian languages while maintaining competitive results on global benchmarks. Designed with focus on reasoning, coding, multilingual capabilities, and agentic tasks. Open-sourced under Apache 2.0 license, optimized for practical deployment from resource-constrained environments to high-performance applications.


其他 Sarvam 模型

评估基准

没有可用的 Sarvam-30B 评估基准。

排名

排名

-

编程排名

-

模型透明度

总分

B

67 / 100

GPU 要求

完整计算器

选择模型权重的量化方法

上下文大小:1024 个令牌

1k
63k
125k

所需显存:

推荐 GPU