Parameters
14B
Context Length
256K
Modality
Multimodal
Architecture
Dense
License
Apache 2.0
Release Date
2 Dec 2025
Knowledge Cutoff
-
Attention Structure
Multi-Head Attention
Hidden Dimension Size
-
Number of Layers
-
Attention Heads
-
Key-Value Heads
-
Activation Function
-
Normalization
-
Position Embedding
Absolute Position Embedding
VRAM requirements for different quantization methods and context sizes
Ministral 3 14B is a sophisticated, dense model within the Ministral 3 family, designed to deliver advanced capabilities while accommodating practical hardware constraints, making it suitable for a range of private AI deployments. It integrates a 13.5 billion parameter language model with a 0.4 billion parameter vision encoder, enabling multimodal understanding and processing of both textual and visual inputs. This model variant is distinguished by its multilingual capabilities, supporting over 40 languages, including major European and East Asian languages.
The architectural design of Ministral 3 14B focuses on efficient performance for edge and local computing environments. It is structured as a dense model, contrasting with Mixture-of-Experts (MoE) architectures found in larger models. The model supports a substantial context window of 256,000 tokens, facilitating the processing of extensive inputs and enabling more comprehensive interactions. Its attention mechanism is a Multi-Head Attention (MHA) structure, a standard component in transformer-based models that allows for processing different parts of the input sequence concurrently.
In terms of functionality, Ministral 3 14B offers advanced agentic features, including native function calling and structured JSON output, which enhances its utility in complex automation and conversational AI systems. This model is optimized for diverse use cases such as private chat applications, local AI assistants, and fine-tuning for specialized tasks. Its design prioritizes performance at a smaller scale, ensuring deployability across various hardware configurations, including those with limited resources.
Ministral 3 is a family of efficient edge models with vision capabilities, available in 3B, 8B, and 14B parameter sizes. Designed for edge deployment with multimodal and multilingual support, offering best-in-class performance for resource-constrained environments.
Ranking is for Local LLMs.
No evaluation benchmarks for Ministral 3 14B available.
Overall Rank
-
Coding Rank
-
Full Calculator
Choose the quantization method for model weights
Context Size: 1,024 tokens