ApX 标志ApX 标志

趋近智

GPT-5.1 Codex Mini

参数

-

上下文长度

400K

模态

Text

架构

Dense

许可证

Proprietary

发布日期

13 Nov 2025

训练数据截止日期

Sep 2024

技术规格

注意力结构

Multi-Head Attention

隐藏维度大小

-

层数

-

注意力头

-

键值头

-

激活函数

-

归一化

-

位置嵌入

Absolute Position Embedding

GPT-5.1 Codex Mini

GPT-5.1 Codex Mini is a specialized, lightweight large language model engineered to facilitate rapid software development and streamlined coding workflows. As a high-efficiency variant within the GPT-5.1 series, it is optimized for low-latency performance in environments requiring immediate feedback, such as real-time code completion, inline refactoring, and interactive debugging within integrated development environments (IDEs). The model is designed to handle routine programming tasks with a focus on high throughput and reduced computational overhead, making it a cost-effective alternative for developers who require consistent assistance without the resource requirements of larger reasoning models.

Technically, the model employs a dense transformer architecture utilizing Multi-Head Attention (MHA) and absolute position embeddings. This design choice ensures predictable and deterministic outputs critical for syntax-heavy tasks where structural accuracy is paramount. It supports a substantial context window of 400,000 tokens, enabling it to ingest large portions of a codebase or extensive documentation for more contextualized generation. The model's training focuses on code-specific datasets, including a vast corpus of multi-language repositories and software documentation, which allows it to maintain precision in logic and syntax across common programming languages like Python, JavaScript, and C++.

Functionally, GPT-5.1 Codex Mini operates as a workhorse for developer-centric applications, supporting advanced features such as function calling, structured outputs, and vision-integrated UI development. It is capable of processing multimodal inputs, specifically interpreting screenshots or design mockups to generate corresponding frontend code or assist in visual debugging. By balancing raw generation speed with reliable instruction following, the model serves as a core component for agentic coding tools and CI/CD pipelines where automated code review and unit test generation are performed at scale.

关于 GPT-5

OpenAI's latest generation of language models featuring advanced reasoning capabilities, extended context windows up to 400K tokens, and specialized variants for coding, general intelligence, and efficiency. GPT-5 series introduces improved thinking modes, superior performance across benchmarks, and variants optimized for different use cases from high-capacity Pro models to efficient Nano models. Features native multimodal understanding, enhanced mathematical reasoning, and state-of-the-art coding abilities through Codex variants.


其他 GPT-5 模型

评估基准

排名

#31

基准分数排名

0.98

🥈

2

0.65

18

Agentic Coding

LiveBench Agentic

0.40

18

0.76

19

0.70

20

0.70

26

排名

排名

#31

编程排名

#34