ApX 标志

趋近智

Ministral 3 3B

参数

3B

上下文长度

256K

模态

Multimodal

架构

Dense

许可证

Apache 2.0

发布日期

2 Dec 2025

训练数据截止日期

-

技术规格

注意力结构

Multi-Head Attention

隐藏维度大小

-

层数

-

注意力头

-

键值头

8

激活函数

-

归一化

Layer Normalization

位置嵌入

Absolute Position Embedding

系统要求

不同量化方法和上下文大小的显存要求

Ministral 3 3B

The Ministral 3 3B model is a compact, multimodal language model developed by Mistral AI, specifically engineered for efficient deployment in resource-constrained environments such as edge devices. It integrates a 3.4 billion parameter language model with a 0.4 billion parameter vision encoder, resulting in a total of 3.8 billion parameters. This architecture facilitates advanced capabilities including multimodal understanding, allowing the model to process and interpret both text and visual inputs. Designed to operate with a minimal memory footprint, Ministral 3 3B is capable of running locally on devices with limited VRAM, enhancing its applicability for on-device inference and privacy-sensitive applications.

Architecturally, Ministral 3 3B is a dense Transformer model that incorporates Grouped Query Attention (GQA) to optimize processing speed and memory utilization. This attention mechanism contributes to the model's ability to efficiently handle long input sequences, supporting a context length of up to 256,000 tokens. The model's design includes 8 key-value heads, which aid in capturing complex relationships within input data while maintaining computational efficiency. These technical considerations ensure a balance between performance and the practical constraints of edge computing.

The Ministral 3 3B model is suitable for a range of lightweight, real-time applications, including image captioning, text classification, real-time translation, content generation, and data extraction on edge devices. Its inherent multimodal and multilingual capabilities, supporting dozens of languages, further broaden its applicability across diverse use cases requiring local intelligence. The model also offers robust support for agentic workflows, featuring native function calling and structured JSON output, making it effective for orchestrating multi-step tasks and specialized applications.

关于 Ministral 3

Ministral 3 is a family of efficient edge models with vision capabilities, available in 3B, 8B, and 14B parameter sizes. Designed for edge deployment with multimodal and multilingual support, offering best-in-class performance for resource-constrained environments.


其他 Ministral 3 模型

评估基准

排名适用于本地LLM。

没有可用的 Ministral 3 3B 评估基准。

排名

排名

-

编程排名

-

GPU 要求

完整计算器

选择模型权重的量化方法

上下文大小:1024 个令牌

1k
125k
250k

所需显存:

推荐 GPU