Instruction Following Fine-Tuning using Generated Data
New · Open Source
Kerb - LLM Development Toolkit
Python toolkit for building production-ready LLM applications. Modular utilities for prompts, RAG, agents, structured outputs, and multi-provider support.
Was this section helpful?
Self-Instruct: Aligning Language Models with Self-Generated Instructions, Yizhong Wang, Yeganeh Kordi, Swaroop Mishra, Alisa Liu, Noah A. Smith, Daniel Khashabi, Hannaneh Hajishirzi, 2022ACL 2023 (Association for Computational Linguistics)DOI: 10.48550/arXiv.2212.10560 - A foundational paper introducing a method for large language models to generate their own instruction-following data, reducing the need for extensive human annotation.
Finetuned Language Models Are Zero-Shot Learners, Jason Wei, Maarten Bosma, Vincent Y. Zhao, Kelvin Guu, Adams Wei Yu, Brian Lester, Nan Du, Andrew M. Dai, and Quoc V. Le, 2022arXiv preprint arXiv:2109.01652DOI: 10.48550/arXiv.2109.01652 - Introduces instruction tuning as a method to improve the zero-shot performance of language models across a wide range of tasks, laying groundwork for instruction-following models.