Quantization and its effect on Fine-Tuning (QLoRA)
New ยท Open Source
Kerb - LLM Development Toolkit
Python toolkit for building production-ready LLM applications. Modular utilities for prompts, RAG, agents, structured outputs, and multi-provider support.
LoRA: Low-Rank Adaptation of Large Language Models, Edward J. Hu, Yelong Shen, Phillip Wallis, Zeyuan Allen-Zhu, Yuanzhi Li, Shean Wang, Lu Wang, Weizhu Chen, 2021arXiv preprint arXiv:2106.09685DOI: 10.48550/arXiv.2106.09685 - Presents the foundational LoRA technique for parameter-efficient adaptation of pre-trained models.
bitsandbytes GitHub Repository, Tim Dettmers and collaborators, 2024 - Official repository for the bitsandbytes library, which provides the 4-bit quantization implementation used in QLoRA.