QLoRA: Efficient Finetuning of Quantized LLMs, Tim Dettmers, Artidoro Pagnoni, Ari Holtzman, Luke Zettlemoyer, 2023arXiv preprint arXiv:2305.14314DOI: 10.48550/arXiv.2305.14314 - This paper introduces QLoRA, detailing 4-bit NormalFloat (NF4) quantization, Double Quantization (DQ), and Paged Optimizers. It provides the technical basis for the section.
Parameter-Efficient Fine-Tuning (PEFT) library, Hugging Face, 2024 (Hugging Face) - Official documentation for the Hugging Face PEFT library, which offers high-level APIs for applying LoRA, QLoRA, and other parameter-efficient fine-tuning methods.