Efficient Estimation of Word Representations in Vector Space, Tomas Mikolov, Kai Chen, Greg Corrado, and Jeffrey Dean, 2013DOI: 10.48550/arXiv.1301.3781 - This paper introduces the Word2Vec model, which revolutionized word embedding learning by efficiently producing high-quality dense vector representations that capture semantic and syntactic relationships.
GloVe: Global Vectors for Word Representation, Jeffrey Pennington, Richard Socher, Christopher Manning, 2014Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP) (Association for Computational Linguistics)DOI: 10.3115/v1/D14-1162 - This paper presents GloVe, an alternative word embedding method that combines global matrix factorization and local context window methods to learn vector representations from word-word co-occurrence statistics.