LoRA: Low-Rank Adaptation of Large Language Models, Edward J. Hu, Yelong Shen, Phillip Wallis, Zeyuan Allen-Zhu, Yuanzhi Li, Shean Wang, Lu Wang, Weizhu Chen, 2021International Conference on Learning Representations (ICLR)DOI: 10.48550/arXiv.2106.09685 - Introduces Low-Rank Adaptation (LoRA), a foundational PEFT method, and analyzes its performance and efficiency in adapting large models, including aspects of generalization.
QLoRA: Efficient Finetuning of Quantized LLMs, Tim Dettmers, Artidoro Pagnoni, Ari Holtzman, Luke Zettlemoyer, 2023Advances in Neural Information Processing Systems (NeurIPS)DOI: 10.48550/arXiv.2305.14314 - Presents QLoRA, an efficient fine-tuning method that quantizes a pre-trained LLM and then fine-tunes with LoRA. Discusses its impact on memory and performance, and implicitly on generalization and robustness properties when applied to quantized models.