Attention Is All You Need, Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Łukasz Kaiser, and Illia Polosukhin, 2017Advances in Neural Information Processing Systems 30 (NeurIPS)DOI: 10.5555/3295222.3295349 - 引入Transformer架构的开创性论文,该架构是大多数现代大型语言模型的基础,也是本实践中使用的特定文档。
Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks, Patrick Lewis, Ethan Perez, Aleksandra Piktus, Fabio Petroni, Vladimir Karpukhin, Narsimha Chilkuri, Michael Grave, Pasquale Minervini, and Sebastian Riedel, 2020Advances in Neural Information Processing Systems 33, Vol. 33 (NeurIPS)DOI: 10.5555/3495289.3495444 - 介绍了检索增强生成(RAG)架构,该架构结合了参数化和非参数化记忆,显著提升了知识密集型自然语言处理任务的性能。