Parameter-Efficient Transfer Learning for NLP, Neil Houlsby, Andrei Giurgiu, Stanislaw Jastrzebski, Bruna Morrone, Quentin de Laroussilhe, Andrea Gesmundo, Mona Attariyan, Sylvain Gelly, 2019Proceedings of the 36th International Conference on Machine Learning (ICML)DOI: 10.48550/arXiv.1902.00751 - 介绍了用于高效迁移学习的适配器模块,通过冻结预训练权重并插入小型可训练层来展示其有效性。
Prefix-Tuning: Optimizing Continuous Prompts for Generation, Li, Xiang Lisa and Liang, Percy, 2021Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers) (Association for Computational Linguistics)DOI: 10.18653/v1/2021.acl-long.353 - 提出了一种参数高效的微调方法--Prefix-Tuning,该方法在每个Transformer层注意力机制的输入前添加可训练的连续前缀。