LM Studio Official Website, LM Studio Team, 2024 - Provides the official software, guides, and community resources for LM Studio, essential for downloading, managing, and interacting with local large language models.
llama.cpp GitHub Repository, Georgi Gerganov and the llama.cpp contributors, 2024 - The foundational project for efficient CPU/GPU inference of large language models, including the development of the GGUF file format used by LM Studio.
A Survey of Large Language Models, Wayne Xin Zhao, Kun Zhou, Junyi Li, Tianyi Tang, Xiaolei Wang, Yupeng Hou, Yingqian Min, Beichen Zhang, Junjie Zhang, Zican Dong, Yifan Du, Chen Yang, Yushuo Chen, Zhipeng Chen, Jinhao Jiang, Ruiyang Ren, Yifan Li, Xinyu Tang, Zikang Liu, Peiyu Liu, Jian-Yun Nie, Ji-Rong Wen, 2023arXiv preprint arXiv:2303.18223DOI: 10.48550/arXiv.2303.18223 - Provides a comprehensive overview of large language models, covering their architectures, training, capabilities, and applications, useful for understanding the underlying technology.
QLoRA: Efficient Finetuning of Quantized LLMs, Tim Dettmers, Artidoro Pagnoni, Ari Holtzman, Luke Zettlemoyer, 2023arXiv preprint arXiv:2305.14314DOI: 10.48550/arXiv.2305.14314 - Introduces techniques for efficient finetuning of quantized large language models, providing insights into how quantization reduces memory footprint and impacts performance for local deployment.