Notice: The reproducibility variables underlying each score are classified using an automated LLM-based pipeline, validated against a manually labeled dataset. LLM-based classification introduces uncertainty and potential bias; scores should be interpreted as estimates. Full accuracy metrics and methodology are described in [1].

Modeling Transitivity and Cyclicity in Directed Graphs via Binary Code Box Embeddings

Authors: Dongxu Zhang, Michael Boratko, Cameron Musco, Andrew McCallum

NeurIPS 2022 | Venue PDF | LLM Run Details

Reproducibility Variable Result LLM Response
Research Type Experimental Theoretical and empirical results show that the proposed models not only preserve a useful inductive bias of transitivity but also have suf๏ฌcient representational capacity to model arbitrary graphs, including graphs with cycles. We evaluate our model on graph reconstruction and link prediction tasks with various synthetic graphs and real world graphs, and observe that our proposed methods perform the best in almost all scenarios, especially when a graph has strong cyclicity and transitivity.
Researcher Affiliation Academia University of Massachusetts Amherst EMAIL
Pseudocode No The paper includes mathematical formulations and proofs but no distinct pseudocode or algorithm blocks.
Open Source Code Yes 1Our code is available at https://github.com/iesl/geometric_graph_embedding.
Open Datasets Yes We evaluate on the following real world graphs: Google hyperlink network, Epinions trust network, CORA citation network, Twitter network, and DREAM5 gene regulatory networks. For more data statistics, see Appendix E. We obtained code and data for real world graphs that was publicly available.
Dataset Splits Yes During training, all hyperparameters are tuned via 10% cross-validation and the test set results are averaged over 10 models of different random seeds, which were trained on the full training set with the best hyper-parameters found during crossvalidation.
Hardware Specification No The paper does not provide specific details about the hardware used for experiments. The ethics checklist states: '3. If you ran experiments... (d) Did you include the total amount of compute and the type of resources used (e.g., type of GPUs, internal cluster, or cloud provider)? [No]'
Software Dependencies No The paper does not list specific versions for software dependencies (e.g., Python, PyTorch versions).
Experiment Setup Yes We use Bayesian hyperparameter tuning based on the optimal F1 score for reconstruction. We sample minibatches of positive edges in T , and for each positive edge we sample 32 edges not in T by randomly corrupting either the source or target node. We also use a self-adversarial negative weight, as described in [26]. During training, all hyperparameters are tuned via 10% cross-validation. For fair comparison, we follow [30] and use 128 parameters per vertex for all our models. We also tune the weight of the sparsity regularization wr for all BC-Box models (see Section 4.5).