趋近智
参数
104B
上下文长度
128K
模态
Text
架构
Dense
许可证
CC-BY-NC
发布日期
4 Apr 2024
知识截止
Feb 2023
注意力结构
Multi-Head Attention
隐藏维度大小
-
层数
-
注意力头
-
键值头
-
激活函数
-
归一化
-
位置嵌入
Absolute Position Embedding
不同量化方法和上下文大小的显存要求
Cohere Command R Plus is a large language model developed by Cohere, designed to support demanding enterprise applications. This model is optimized for conversational interactions and tasks requiring extensive context, such as advanced Retrieval Augmented Generation (RAG) and multi-step tool use. Its primary function is to enable organizations to deploy sophisticated AI capabilities in production environments by balancing computational efficiency with high accuracy. The model offers comprehensive multilingual support, having been trained and evaluated across ten key global business languages, with additional pre-training data from thirteen other languages, facilitating its applicability in diverse linguistic contexts.
From an architectural standpoint, Command R Plus utilizes an optimized transformer architecture. Following its pretraining phase, the model undergoes supervised fine-tuning and preference training processes. These alignment procedures are crucial for refining the model's behavior to meet human preferences for helpfulness and safety in its generative outputs. A notable technical characteristic is its support for an expansive context window, capable of processing up to 128,000 tokens. This significant context length is achieved through specialized design, incorporating innovations in positional encodings to manage long dependencies effectively.
In terms of operational performance and applications, Command R Plus is engineered for large-scale production workloads. It demonstrates proficiency in complex RAG workflows, delivering grounded responses with inline citations to enhance reliability and mitigate issues such as hallucination. The model's multi-step tool use capabilities enable it to automate intricate business processes, including structured data analysis and dynamic updates within systems like customer relationship management (CRM). Furthermore, it possesses the ability to perform self-correction in instances of tool failure, thereby increasing the overall success rate of automated tasks. An August 2024 update to the model introduced enhancements, resulting in approximately 50% higher throughput and 25% lower latencies while maintaining the same hardware footprint.
排名适用于本地LLM。
排名
#46
基准 | 分数 | 排名 |
---|---|---|
General Knowledge MMLU | 0.69 | 10 |
Refactoring Aider Refactoring | 0.38 | 12 |
Coding Aider Coding | 0.38 | 15 |
Agentic Coding LiveBench Agentic | 0.02 | 19 |
Data Analysis LiveBench Data Analysis | 0.49 | 23 |
Coding LiveBench Coding | 0.27 | 26 |
Reasoning LiveBench Reasoning | 0.22 | 27 |
Mathematics LiveBench Mathematics | 0.23 | 30 |