Graph Attention Networks, Petar Veličković, Guillem Cucurull, Arantxa Casanova, Adriana Romero, Pietro Liò, Yoshua Bengio, 2018International Conference on Learning Representations (ICLR)DOI: 10.48550/arXiv.1710.10903 - The original paper introducing Graph Attention Networks, detailing the attention mechanism and multi-head attention.
Attention Is All You Need, Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Lukasz Kaiser, Illia Polosukhin, 2017Advances in Neural Information Processing Systems (NeurIPS)DOI: 10.48550/arXiv.1706.03762 - This foundational paper introduced the Transformer architecture and the self-attention mechanism, which inspired GATs' attention approach.