Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks, Nils Reimers and Iryna Gurevych, 2019Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP) (Association for Computational Linguistics)DOI: 10.18653/v1/D19-1410 - 本文介绍了Sentence-BERT,一个广泛使用的生成句子嵌入的框架,也是Sentence Transformers库的基础,与讨论的基础模型和微调策略直接相关。
LoRA: Low-Rank Adaptation of Large Language Models, Edward J. Hu, Yelong Shen, Phillip Wallis, Zeyuan Allen-Zhu, Yuanzhi Li, Shean Wang, Lu Wang, and Weizhu Chen, 2022International Conference on Learning Representations (ICLR 2022)DOI: 10.48550/arXiv.2106.09685 - 这篇论文介绍了低秩适应(LoRA),一种参数高效的微调方法,它解决了计算成本和灾难性遗忘问题,正如文中对基于适配器的微调所描述的。