Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks, Patrick Lewis, Ethan Perez, Aleksandra Piktus, Fabio Petroni, Vladimir Karpukhin, Naman Goyal, Heinrich Küttler, Mike Lewis, Wen-tau Yih, Tim Rocktäschel, Sebastian Riedel, Douwe Kiela, 2020Advances in Neural Information Processing Systems 33 (NeurIPS 2020), Vol. 33 (Curran Associates, Inc.)DOI: 10.48550/arXiv.2005.11401 - 阐述了最初的检索增强生成(RAG)框架,详细介绍了其架构以及将外部知识融入语言模型的优势。
Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks, Nils Reimers and Iryna Gurevych, 2019Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP) (Association for Computational Linguistics)DOI: 10.18653/v1/D19-1410 - 介绍了Sentence-BERT模型,它能高效生成高质量的句子嵌入,是检索器中使用的sentence-transformers等库的基础。