Continuous Hierarchical Representations with Poincaré Variational Auto-Encoders
Authors: Emile Mathieu, Charline Le Lan, Chris J. Maddison, Ryota Tomioka, Yee Whye Teh
NeurIPS 2019 | Conference PDF | Archive PDF | Plain Text | LLM Run Details
| Reproducibility Variable | Result | LLM Response |
|---|---|---|
| Research Type | Experimental | We empirically show better generalisation to unseen data than the Euclidean counterpart, and can qualitatively and quantitatively better recover hierarchical structures. (Abstract) ... We implemented our model and ran our experiments within the automatic differentiation framework Py Torch (Paszke et al., 2017). ... 5 Experiments |
| Researcher Affiliation | Collaboration | Emile Mathieu emile.mathieu@stats.ox.ac.uk, Department of Statistics, University of Oxford, United Kingdom; Chris J. Maddison cmaddis@stats.ox.ac.uk, Deep Mind, London, United Kingdom; Ryota Tomioka ryoto@microsoft.com, Microsoft Research, Cambridge, United Kingdom |
| Pseudocode | Yes | Algorithm 1 Hyperbolic normal sampling scheme |
| Open Source Code | Yes | We open-source our code for reproducibility and to benefit the community 1. 1https://github.com/emilemathieu/pvae |
| Open Datasets | Yes | We assess our modelling assumption on data generated from a branching diffusion process... The MNIST (Le Cun and Cortes, 2010) dataset... We demonstrate these capabilities on three network datasets: a graph of Ph.D. advisor-advisee relationships (Nooy et al., 2011), a phylogenetic tree expressing genetic heritage (Hofbauer et al., 2016; Sanderson and Eriksson, 1994) and a biological set representing disease relationships (Goh et al., 2007; Rossi and Ahmed, 2015). |
| Dataset Splits | No | The paper mentions 'Learning rates are chosen by cross-validation among {0.01, 0.005, 0.001}' in Appendix C, implying a validation process for hyperparameter tuning. However, it does not explicitly provide specific percentages or sample counts for a dedicated validation dataset split, nor does it explicitly mention a 'validation set' in the main text with its dimensions. |
| Hardware Specification | No | The paper states, 'We implemented our model and ran our experiments within the automatic differentiation framework Py Torch (Paszke et al., 2017).', but it does not specify any details about the hardware used, such as CPU or GPU models, memory, or specific computing environments. |
| Software Dependencies | No | The paper mentions using 'Py Torch (Paszke et al., 2017)' and the 'Adam optimizer (Kingma and Ba, 2016)'. However, it does not provide specific version numbers for these software components (e.g., 'PyTorch 1.9' or 'Adam version X'), which are necessary for full reproducibility. |
| Experiment Setup | Yes | Experimental details are fully described in Appendix C. ... We train all models for 100 epochs using the Adam optimizer (Kingma and Ba, 2016) with a batch size of 100 for the synthetic data and 500 for MNIST. For the graph datasets we use a batch size of 1 for the node features and 200 for the graph edges. Learning rates are chosen by cross-validation among {0.01, 0.005, 0.001} and the prior dispersion for the P-VAE among {0.1, 0.3, 0.8, 1.0, 1.2, 1.4}. |