Instruction Following Fine-Tuning using Generated Data
Was this section helpful?
Self-Instruct: Aligning Language Models with Self-Generated Instructions, Yizhong Wang, Yeganeh Kordi, Swaroop Mishra, Alisa Liu, Noah A. Smith, Daniel Khashabi, Hannaneh Hajishirzi, 2022ACL 2023 (Association for Computational Linguistics)DOI: 10.48550/arXiv.2212.10560 - A foundational paper introducing a method for large language models to generate their own instruction-following data, reducing the need for extensive human annotation.
Finetuned Language Models Are Zero-Shot Learners, Jason Wei, Maarten Bosma, Vincent Y. Zhao, Kelvin Guu, Adams Wei Yu, Brian Lester, Nan Du, Andrew M. Dai, and Quoc V. Le, 2022arXiv preprint arXiv:2109.01652DOI: 10.48550/arXiv.2109.01652 - Introduces instruction tuning as a method to improve the zero-shot performance of language models across a wide range of tasks, laying groundwork for instruction-following models.