Attention Is All You Need, Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Łukasz Kaiser, Illia Polosukhin, 2017Advances in Neural Information Processing Systems 30 (Curran Associates, Inc.)DOI: 10.5591/978-1-57766-068-1.5998 - This introduces the Transformer architecture, self-attention, and multi-head attention, forming the basis for subsequent models in sequence processing.
Conformer: Convolution-augmented Transformer for Speech Recognition, Anmol Gulati, James Qin, C. C. Chiu, Haikal Hassoun, Ruoming Pang, Karthik Ramasubramanian, Justin Salpekar, Kevin Webster, Yonghui Wu, 2020INTERSPEECH 2020 (International Speech Communication Association (ISCA))DOI: 10.21437/Interspeech.2020-1168 - This presents the Conformer architecture, which integrates convolutions into Transformer blocks for improved speech recognition performance.