Continuous Hierarchical Representations with Poincaré Variational Auto-Encoders

Authors: Emile Mathieu, Charline Le Lan, Chris J. Maddison, Ryota Tomioka, Yee Whye Teh

NeurIPS 2019 | Conference PDF | Archive PDF | Plain Text | LLM Run Details

Reproducibility Variable Result LLM Response
Research Type Experimental We empirically show better generalisation to unseen data than the Euclidean counterpart, and can qualitatively and quantitatively better recover hierarchical structures. (Abstract) ... We implemented our model and ran our experiments within the automatic differentiation framework Py Torch (Paszke et al., 2017). ... 5 Experiments
Researcher Affiliation Collaboration Emile Mathieu emile.mathieu@stats.ox.ac.uk, Department of Statistics, University of Oxford, United Kingdom; Chris J. Maddison cmaddis@stats.ox.ac.uk, Deep Mind, London, United Kingdom; Ryota Tomioka ryoto@microsoft.com, Microsoft Research, Cambridge, United Kingdom
Pseudocode Yes Algorithm 1 Hyperbolic normal sampling scheme
Open Source Code Yes We open-source our code for reproducibility and to benefit the community 1. 1https://github.com/emilemathieu/pvae
Open Datasets Yes We assess our modelling assumption on data generated from a branching diffusion process... The MNIST (Le Cun and Cortes, 2010) dataset... We demonstrate these capabilities on three network datasets: a graph of Ph.D. advisor-advisee relationships (Nooy et al., 2011), a phylogenetic tree expressing genetic heritage (Hofbauer et al., 2016; Sanderson and Eriksson, 1994) and a biological set representing disease relationships (Goh et al., 2007; Rossi and Ahmed, 2015).
Dataset Splits No The paper mentions 'Learning rates are chosen by cross-validation among {0.01, 0.005, 0.001}' in Appendix C, implying a validation process for hyperparameter tuning. However, it does not explicitly provide specific percentages or sample counts for a dedicated validation dataset split, nor does it explicitly mention a 'validation set' in the main text with its dimensions.
Hardware Specification No The paper states, 'We implemented our model and ran our experiments within the automatic differentiation framework Py Torch (Paszke et al., 2017).', but it does not specify any details about the hardware used, such as CPU or GPU models, memory, or specific computing environments.
Software Dependencies No The paper mentions using 'Py Torch (Paszke et al., 2017)' and the 'Adam optimizer (Kingma and Ba, 2016)'. However, it does not provide specific version numbers for these software components (e.g., 'PyTorch 1.9' or 'Adam version X'), which are necessary for full reproducibility.
Experiment Setup Yes Experimental details are fully described in Appendix C. ... We train all models for 100 epochs using the Adam optimizer (Kingma and Ba, 2016) with a batch size of 100 for the synthetic data and 500 for MNIST. For the graph datasets we use a batch size of 1 for the node features and 200 for the graph edges. Learning rates are chosen by cross-validation among {0.01, 0.005, 0.001} and the prior dispersion for the P-VAE among {0.1, 0.3, 0.8, 1.0, 1.2, 1.4}.