Was this section helpful?
transformers
library, detailing how to load models with 4-bit and 8-bit quantization.optimum
library, which enables model optimization, quantization, and deployment on various hardware.bitsandbytes
: 8-bit Optimizers and Quantization Functions for PyTorch, Tim Dettmers, 2023 - The GitHub repository for the bitsandbytes
library, providing core 4-bit and 8-bit quantization functionalities for PyTorch models.