趋近智
参数
300M
上下文长度
131.072K
模态
Text
架构
Dense
许可证
Apache 2.0
发布日期
30 Jun 2025
知识截止
-
注意力结构
Multi-Head Attention
隐藏维度大小
-
层数
18
注意力头
16
键值头
2
激活函数
-
归一化
-
位置嵌入
Absolute Position Embedding
不同量化方法和上下文大小的显存要求
The ERNIE-4.5-0.3B model represents a foundational component within Baidu's ERNIE 4.5 model family. This model is designed as a compact and efficient language model, focusing on general-purpose natural language understanding and generation tasks. Unlike the larger Mixture-of-Experts (MoE) variants in the ERNIE 4.5 series, ERNIE-4.5-0.3B employs a dense transformer architecture, making it suitable for high-throughput applications and resource-constrained environments. Its engineering prioritizes robust language capabilities while maintaining a minimal operational footprint, enabling deployment across diverse scenarios where computational efficiency is paramount.
Architecturally, ERNIE-4.5-0.3B is built upon a dense transformer framework, featuring 18 layers and 16 attention heads. This configuration supports its text processing capabilities, facilitating the interpretation and generation of textual content. The model benefits from training methodologies shared across the broader ERNIE 4.5 family, which incorporate advanced techniques for optimized efficiency. While larger, multimodal ERNIE 4.5 models integrate heterogeneous Mixture-of-Experts and specific multimodal training, ERNIE-4.5-0.3B maintains its focus on delivering high performance within its text-only, dense design.
Designed for practical application, ERNIE-4.5-0.3B is particularly well-suited for high-throughput and edge computing deployments. Its use cases encompass tasks such as sentiment analysis, topic categorization, and spam detection at scale. The model's optimized design further enables direct execution on mobile devices, supporting real-time applications such as text completion or streamlined question-answering systems. Additionally, ERNIE-4.5-0.3B is capable of generating concise summaries of text documents, providing an efficient tool for rapid information review.
The Baidu ERNIE 4.5 family consists of ten large-scale multimodal models. They utilize a heterogeneous Mixture-of-Experts (MoE) architecture, which enables parameter sharing across modalities while also employing dedicated parameters for specific modalities, supporting efficient language and multimodal processing.
排名适用于本地LLM。
没有可用的 ERNIE-4.5-0.3B 评估基准。