Parameter-Efficient Transfer Learning for NLP, Neil Houlsby, Andrei Giurgiu, Stanislaw Jastrzebski, Bruna Morrone, Quentin de Laroussilhe, Andrea Gesmundo, Mona Attariyan, Sylvain Gelly, 2019Proceedings of the 36th International Conference on Machine Learning (ICML)DOI: 10.48550/arXiv.1902.00751 - Introduces adapter modules for efficient transfer learning, demonstrating their effectiveness by freezing pre-trained weights and inserting small, trainable layers.
Prefix-Tuning: Optimizing Continuous Prompts for Generation, Li, Xiang Lisa and Liang, Percy, 2021Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers) (Association for Computational Linguistics)DOI: 10.18653/v1/2021.acl-long.353 - Proposes prefix-tuning, a parameter-efficient fine-tuning method that prepends trainable continuous prefixes to the input of each transformer layer's attention mechanism.