LoRA: Low-Rank Adaptation of Large Language Models, Edward J. Hu, Yelong Shen, Phillip Wallis, Zeyuan Allen-Zhu, Yuanzhi Li, Shean Wang, Lu Wang, Weizhu Chen, 2021International Conference on Learning Representations (ICLR)DOI: 10.48550/arXiv.2106.09685 - Presents the foundational technique of Low-Rank Adaptation for efficient fine-tuning of large language models.
QLoRA: Efficient Finetuning of Quantized LLMs, Tim Dettmers, Artidoro Pagnoni, Ari Holtzman, Luke Zettlemoyer, 2023arXiv preprint arXiv:2305.14314DOI: 10.48550/arXiv.2305.14314 - Introduces QLoRA, enabling effective fine-tuning of 4-bit quantized large language models with reduced memory consumption.
Parameter-Efficient Fine-tuning (PEFT) library, Hugging Face, 2024 (Hugging Face) - Official documentation for the Hugging Face PEFT library, offering practical guides and API details for implementing efficient fine-tuning methods like LoRA.