Parameters
8B
Context Length
128K
Modality
Text
Architecture
Dense
License
Apache-2.0
Release Date
1 Jun 2024
Knowledge Cutoff
-
Attention Structure
Multi-Head Attention
Hidden Dimension Size
-
Number of Layers
-
Attention Heads
-
Key-Value Heads
-
Activation Function
-
Normalization
-
Position Embedding
Absolute Position Embedding
VRAM requirements for different quantization methods and context sizes
Typhoon-2-8B is an 8 billion parameter model optimized for Thai language processing. It features an expanded context length of 128,000 tokens and supports function calling. The model is trained to handle Thai cultural nuances and specific domains such as Thai law and local administration. Released under the Apache 2.0 license.
Typhoon is a Thai language model family developed by SCB 10X. It is specifically optimized for the Thai language, addressing complexities such as the lack of word delimiters and tonal nuances. The models are trained on Thai-centric datasets including legal, cultural, and historical documents to ensure localized context and knowledge.
No evaluation benchmarks for Typhoon-2-8B available.
Overall Rank
-
Coding Rank
-
Full Calculator
Choose the quantization method for model weights
Context Size: 1,024 tokens