BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding, Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova, 2019Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers) (Association for Computational Linguistics)DOI: 10.18653/v1/N19-1423 - 这篇里程碑式的论文提出了一种强大的语言模型预训练技术和成功的微调策略,成为许多后续大型语言模型的标准。
Transfer Learning in Natural Language Processing, Sebastian Ruder, Matthew E. Peters, Swabha Swayamdipta, Thomas Wolf, 2019Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Tutorials (Association for Computational Linguistics)DOI: 10.18653/v1/N19-5004 - 本教程全面概述了应用于自然语言处理的迁移学习技术,是理解其在大型语言模型中原理和应用的优质资源。