Notice: The reproducibility variables underlying each score are classified using an automated LLM-based pipeline, validated against a manually labeled dataset. LLM-based classification introduces uncertainty and potential bias; scores should be interpreted as estimates. Full accuracy metrics and methodology are described in [1].

Boosting Graph Neural Networks via Adaptive Knowledge Distillation

Authors: Zhichun Guo, Chunhui Zhang, Yujie Fan, Yijun Tian, Chuxu Zhang, Nitesh V. Chawla

AAAI 2023 | Venue PDF | LLM Run Details

Reproducibility Variable Result LLM Response
Research Type Experimental Extensive experiments have demonstrated the effectiveness of BGNN. In particular, we achieve up to 3.05% improvement for node classification and 6.35% improvement for graph classification over vanilla GNNs. We conduct extensive experimental studies on both tasks, and the results demonstrate the superior performance of BGNN compared with a set of baseline methods.
Researcher Affiliation Academia 1University of Notre Dame, Notre Dame, IN 46556 2Brandeis University, Waltham, MA 02453 3Case Western Reserve University, Cleveland, OH 44106
Pseudocode No The paper describes the model and its components in text and mathematical formulas but does not include any pseudocode or algorithm blocks.
Open Source Code No The paper does not provide any explicit statements about releasing source code or links to a code repository for the described methodology.
Open Datasets Yes We use seven datasets to conduct graph classification and node classification experiments. We follow the data split as in original papers (Sen et al. 2008; Namata et al. 2012) for Cora, Citeseer and Pubmed while the remaining datasets are randomly split using an empirical ratio. The more details are in Section B of Appendix.
Dataset Splits Yes We follow the data split as in original papers (Sen et al. 2008; Namata et al. 2012) for Cora, Citeseer and Pubmed while the remaining datasets are randomly split using an empirical ratio. The more details are in Section B of Appendix.
Hardware Specification No The paper mentions 'Implementation and experiment details are shown in Section C of Appendix' but does not provide any specific hardware details in the main text.
Software Dependencies No The paper mentions 'Implementation and experiment details are shown in Section C of Appendix' but does not provide specific software dependencies with version numbers in the main text.
Experiment Setup Yes For all the GNN backbones, we use two-layer models. For the single teacher setting, we use one GNN as the teacher and a different GNN as the student. For the multiple teacher setting, we permute the three GNNs in six orders, where the first two serve as teachers and the third one is the student. During the training, we clamp the adaptive temperature in the range from 1 to 4, which is a commonly used temperature range in KD.