Parameter-Efficient Transfer Learning for NLP, Neil Houlsby, Andrei Giurgiu, Stanislaw Jastrzebski, Bruna Morrone, Quentin de Laroussilhe, Andrea Gesmundo, Mona Attariyan, Sylvain Gelly, 2019Proceedings of the 36th International Conference on Machine LearningDOI: 10.48550/arXiv.1902.00751 - This work proposes adapter modules as a parameter-efficient fine-tuning method. It provides a technique for implementing multi-task learning by adding task-specific layers while sharing the base model's parameters.
GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding, Alex Wang, Amanpreet Singh, Julian Michael, Felix Hill, Omer Levy, Samuel R. Bowman, 2019ICLR 2019DOI: 10.48550/arXiv.1804.07461 - This paper describes the GLUE benchmark, which profoundly influenced the development and evaluation of multi-task and transfer learning models in natural language processing by offering a diverse set of tasks to assess general language understanding.