Attention Is All You Need, Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Łukasz Kaiser, Illia Polosukhin, 2017Advances in Neural Information Processing Systems 30, Vol. 30 (Neural Information Processing Systems Foundation, Inc. (NeurIPS))DOI: 10.5555/3295222.3295349 - 提出了Transformer架构,它已成为许多现代嵌入模型的基础,实现了上下文文本表示的学习。
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding, Jacob Devlin, Ming-Wei Chang, Kenton Lee, Kristina Toutanova, 2019Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers) (Association for Computational Linguistics)DOI: 10.18653/v1/N19-1423 - 介绍了BERT,这是基于Transformer的预训练语言模型的一项重要进展,对于生成上下文感知的嵌入至关重要。
Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks, Nils Reimers and Iryna Gurevych, 2019Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP) (Association for Computational Linguistics)DOI: 10.18653/v1/D19-1410 - 描述了Sentence-BERT,一种从预训练的类BERT模型创建语义化句子和文档嵌入的方法,并针对语义相似性搜索等任务进行了优化。