趋近智
参数
70B
上下文长度
128K
模态
Text
架构
Dense
许可证
Apache-2.0
发布日期
1 Jun 2024
训练数据截止日期
-
注意力结构
Multi-Head Attention
隐藏维度大小
-
层数
-
注意力头
-
键值头
-
激活函数
-
归一化
-
位置嵌入
Absolute Position Embedding
不同量化方法和上下文大小的显存要求
Typhoon-2-70B is a 70 billion parameter Thai language model with enhanced reasoning and context understanding. It is optimized for complex Thai language tasks including legal document analysis, cultural content generation, and multi-turn conversations. The model maintains the 128K context length of its smaller variant. Released under the Apache 2.0 license.
Typhoon is a Thai language model family developed by SCB 10X. It is specifically optimized for the Thai language, addressing complexities such as the lack of word delimiters and tonal nuances. The models are trained on Thai-centric datasets including legal, cultural, and historical documents to ensure localized context and knowledge.
没有可用的 Typhoon-2-70B 评估基准。