How Attentive are Graph Attention Networks?

Authors: Shaked Brody, Uri Alon, Eran Yahav

ICLR 2022 | Conference PDF | Archive PDF | Plain Text | LLM Run Details

Reproducibility Variable Result LLM Response
Research Type Experimental We perform an extensive evaluation and show that GATv2 outperforms GAT across 12 OGB and other benchmarks while we match their parametric costs. Our code is available at https://github.com/tech-srl/how_attentive_are_ gats.1
Researcher Affiliation Academia Shaked Brody Technion shakedbr@cs.technion.ac.il Uri Alon Language Technologies Institute Carnegie Mellon University ualon@cs.cmu.edu Eran Yahav Technion yahave@cs.technion.ac.il
Pseudocode No The paper describes mathematical formulations for GAT and GATv2 (Equations 6 and 7) but does not provide pseudocode or a clearly labeled algorithm block.
Open Source Code Yes Our code is available at https://github.com/tech-srl/how_attentive_are_ gats.1
Open Datasets Yes We used the provided splits of OGB (Hu et al., 2020) and the Adam optimizer. We further compare GATv2, GAT, and other GNNs on four node-prediction datasets from OGB.
Dataset Splits Yes We tuned hyperparameters according to validation score and early stopping.
Hardware Specification Yes Our experiments consumed approximately 100 days of GPU in total. We used cloud GPUs of type V100, and we used RTX 3080 and 3090 in local GPU machines.
Software Dependencies No GATv2 is available as part of the Py Torch Geometric library, the Deep Graph Library, and the Tensor Flow GNN library. While these libraries are mentioned, no specific version numbers are provided for them or any other software dependencies.
Experiment Setup Yes We tuned the following hyperparameters: number of layers {2, 3, 6}, hidden size {64, 128, 256}, learning rate {0.0005, 0.001, 0.005, 0.01} and sampling method full batch, Graph SAINT (Zeng et al., 2019) and Neighbor Sampling (Hamilton et al., 2017).