Finetuned Language Models are Zero-Shot Learners, Jason Wei, Maarten Bosma, Vincent Y. Zhao, Kelvin Guu, Adams Wei Yu, Brian Lester, Nan Du, Andrew M. Dai, Quoc V. Le, 2022International Conference on Learning Representations (ICLR)DOI: 10.48550/arXiv.2109.01652 - Introduces instruction tuning and shows its effectiveness with a multi-task dataset (FLAN).
Alpaca: A Strong Open-Source Instruction-Following Model, Rohan Taori, Ishaan Gulrajani, Tianyi Zhang, Yann Dubois, Xuechen Li, Carlos Guestrin, Percy Liang, Tatsunori B. Hashimoto, 2023 (Stanford CRFM) - Describes the Alpaca dataset, generated with the Self-Instruct method for instruction-following models.
OpenAssistant Conversations - A New Dataset for Open-Source Instruction Tuning, Andreas Köpf, Yannic Kilcher, Dimitri von Rütte, Sotiris Anagnostidis, Zhi-Rui Tam, Keith Stevens, Abdullah Barhoum, Nguyen Minh Duc, Oliver Stanley, Richárd Nagyfi, Shahul ES, Sameer Suri, David Glushkov, Arnav Dantuluri, Andrew Maguire, Christoph Schuhmann, Huu Nguyen, Alexander Mattick, 2023NeurIPS 2023 Datasets and BenchmarksDOI: 10.48550/arXiv.2304.07327 - Presents the OpenAssistant Conversations dataset, a large crowdsourced dataset for multi-turn instruction tuning.