Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks, Patrick Lewis, Ethan Perez, Aleksandra Piktus, Fabio Petroni, Vladimir Karpukhin, Naman Goyal, Heinrich Küttler, Mike Lewis, Wen-tau Yih, Tim Rocktäschel, Sebastian Riedel, Douwe Kiela, 2020NeurIPS 2020DOI: 10.48550/arXiv.2005.11401 - Introduces the RAG framework, combining retrieval with text generation, laying the groundwork for how external knowledge bases enhance LLM responses.
Lost in the Middle: How Language Models Use Long Contexts, Nelson F. Liu, Kevin Lin, John Hewitt, Ashwin Paranjape, Michele Bevilacqua, Fabio Petroni, Percy Liang, 2023Transactions of the Association for Computational Linguistics (TACL)DOI: 10.48550/arXiv.2307.03172 - Explores how the position of information within an LLM's context window affects its ability to retrieve facts, highlighting the need for context compression and effective placement.