Strategies for Merging Adapters with the Base Model
Was this section helpful?
LoRA: Low-Rank Adaptation of Large Language Models, Edward J. Hu, Yelong Shen, Phillip Wallis, Zeyuan Allen-Zhu, Yuanzhi Li, Shean Wang, Lu Wang, Weizhu Chen, 2021arXiv preprint arXiv:2106.09685DOI: 10.48550/arXiv.2106.09685 - Presents the original LoRA method, detailing its mathematical formulation and parameter efficiency, which underpins the merging strategy.
Hugging Face PEFT Library Documentation, Hugging Face, 2024 (Hugging Face) - Provides official documentation for the Hugging Face PEFT library, including practical guidance on using the merge_and_unload() function for combining adapter weights.
Hugging Face Transformers Library Documentation, Hugging Face, 2024 (Hugging Face) - Official documentation for the Hugging Face transformers library, explaining how to save and load standard models, which applies to models after PEFT adapters are merged.
The Hugging Face Course, Hugging Face, 2024 (Hugging Face) - A comprehensive online course from Hugging Face covering various aspects of training, fine-tuning, and deploying large language models, offering context for production readiness.