Doubly Stochastic Variational Inference for Neural Processes with Hierarchical Latent Variables
Authors: Qi Wang, Herke Van Hoof
ICML 2020 | Conference PDF | Archive PDF | Plain Text | LLM Run Details
| Reproducibility Variable | Result | LLM Response |
|---|---|---|
| Research Type | Experimental | We evaluate this model in several experiments, and our results demonstrate competitive prediction performance in multi-output regression and uncertainty estimation in classification. In this section, we start with learning predictive functions on several toy dataset, and then high-dimensional tasks, including system identification on physics engines, multioutput regression on real-world dataset as well as image classification with uncertainty quantification, are performed to evaluate properties of NP related models. |
| Researcher Affiliation | Academia | 1Amsterdam Machine Learning Lab, University of Amsterdam, Amsterdam, the Netherlands. Correspondence to: Qi Wang <q.wang3@uva.nl>. |
| Pseudocode | Yes | Algorithm 1 Variational Inference for DSVNP in Training. |
| Open Source Code | No | The paper does not provide an explicit statement or link for open-source code for the described methodology. |
| Open Datasets | Yes | We evaluate the performance of all models on dataset, including SARCOS 2, Water Quality (WQ) (Dˇzeroski et al., 2000) and SCM20d (Spyromitros Xioufis et al., 2016). Details about these dataset and neural architectures for all models are included in Appendix E. We respectively train models on MNIST and CIFAR10. |
| Dataset Splits | No | The paper mentions splitting datasets into training and testing sets (e.g., 'Each dataset is randomly split into 2-folds as training and testing sets.'), but does not provide specific details on a separate validation set split or its size/proportion. |
| Hardware Specification | Yes | Also, we gratefully acknowledge the support of NVIDIA Corporation with the donation of a Titan V GPU. |
| Software Dependencies | No | The paper mentions implementation details are in Appendix E, but within the provided text, it does not specify concrete software dependencies with version numbers (e.g., library or solver names with version numbers). |
| Experiment Setup | Yes | The training process follows Algorithm (1) with the maximum number of context points as 100. The dimensions for latent variables are 64 on MNIST and 128 on CIFAR10. The training procedure in (C)NPs follows that in Algorithm (1), and some context points are randomly selected in batch samples. |