Inductive Representation Learning on Large Graphs, William L. Hamilton, Rex Ying, Jure Leskovec, 2017Advances in Neural Information Processing SystemsDOI: 10.48550/arXiv.1706.02216 - Presents the original GraphSAGE framework for inductive node embedding generation, introducing aggregation functions like mean, max-pooling, and LSTM for learning on large graphs.
Principal Neighbourhood Aggregation for Graph Neural Networks, Gabriele Corso, Luca Cavalleri, Dominique Beaini, Pietro Liò, Petar Veličković, 2020Advances in Neural Information Processing SystemsDOI: 10.48550/arXiv.2004.05718 - Proposes Principal Neighbourhood Aggregation, a GNN layer that combines multiple aggregation functions and degree-based scalers to capture richer neighborhood information and handle diverse node degrees effectively.
Graph Attention Networks, Petar Veličković, Guillem Cucurull, Arantxa Casanova, Adriana Romero, Pietro Liò, Yoshua Bengio, 2017International Conference on Learning RepresentationsDOI: 10.48550/arxiv.1710.10903 - Introduces the Graph Attention Network, a neural network architecture that leverages masked self-attentional layers to learn weighted neighbor contributions, influencing subsequent attention-based GNN variants.