Prerequisites: LLM Fundamentals & Python
Level:
Quantization Principles
Understand the fundamental concepts behind model quantization and its benefits for LLMs.
Post-Training Quantization (PTQ)
Implement various PTQ techniques, including calibration and handling outliers.
Advanced PTQ Methods
Apply advanced PTQ algorithms like GPTQ and understand methods like AWQ.
Quantization-Aware Training (QAT)
Understand the concepts of QAT and how to simulate quantization during training.
Formats and Tooling
Work with common quantization formats (GGUF, GPTQ) and libraries (Hugging Face Optimum, bitsandbytes).
Evaluation and Deployment
Evaluate the performance and accuracy trade-offs of quantized LLMs and understand deployment considerations.