趋近智
参数
-
上下文长度
1,048.576K
模态
Multimodal
架构
Dense
许可证
Proprietary
发布日期
25 Sept 2025
训练数据截止日期
Jan 2025
注意力结构
Multi-Head Attention
隐藏维度大小
-
层数
-
注意力头
-
键值头
-
激活函数
-
归一化
-
位置嵌入
Absolute Position Embedding
Gemini 2.5 Flash Max Thinking (2025-09-25) is a high-performance multimodal model engineered to bridge the gap between lightweight execution and advanced cognitive reasoning. Part of the Gemini 2.5 family, this variant is designed to handle complex multi-step tasks by utilizing a native thinking architecture that exposes the model's internal reasoning process before generating a final response. This version, released in September 2025, incorporates significant improvements in instruction following and agentic tool use, making it particularly effective for long-horizon tasks and automated workflows that require high reliability in reasoning-intensive environments.
Technically, the model employs a dense transformer architecture optimized for throughput and efficiency. Unlike standard large-scale models that may sacrifice speed for depth, this Flash variant maintains low-latency performance while supporting an expansive 1.05-million-token context window. It utilizes Multi-Head Attention (MHA) for sequence processing and Absolute Position Embeddings to manage spatial and temporal relationships across its multimodal inputs, which include text, image, audio, and video. The architecture is specifically tuned to provide thinking transparency, allowing developers to monitor and budget reasoning tokens via API parameters, thereby ensuring explainable outputs without the overhead typical of larger proprietary models.
Functionally, Gemini 2.5 Flash Max Thinking is optimized for developers who require a balance of cost-efficiency and intelligence. Its enhanced post-training enables it to excel in coding, mathematics, and scientific analysis by reducing verbosity and improving the accuracy of its chain-of-thought sequences. The model is integrated into the Google AI ecosystem with robust support for function calling, code execution, and grounding through Google Search. This makes it an ideal choice for high-volume applications such as automated research summarization, complex software engineering agents, and multimodal data processing where both speed and logical depth are required.
Google's advanced multimodal models with native understanding of text, images, audio, and video. Features massive context windows up to 2.1M tokens, max thinking modes for complex reasoning, and optimized variants for different performance/cost tradeoffs. Includes Pro, Flash, and Flash Lite variants with configurable thinking capabilities for transparent reasoning.
排名
#49
| 基准 | 分数 | 排名 |
|---|---|---|
Data Analysis LiveBench Data Analysis | 0.73 | 9 |
Mathematics LiveBench Mathematics | 0.75 | 21 |
Reasoning LiveBench Reasoning | 0.51 | 25 |
Coding LiveBench Coding | 0.68 | 32 |
Agentic Coding LiveBench Agentic | 0.23 | 33 |