Partially Stochastic Infinitely Deep Bayesian Neural Networks

Authors: Sergio Calvo OrdoƱez, Matthieu Meunier, Francesco Piatti, Yuantao Shi

ICML 2024 | Conference PDF | Archive PDF | Plain Text | LLM Run Details

Reproducibility Variable Result LLM Response
Research Type Experimental Lastly, empirical evaluations across multiple tasks show that our proposed architectures achieve better downstream task performance and uncertainty quantification than their counterparts while being significantly more efficient. In this section, we detail our Partially Stochastic Infinitely Deep BNNs performance for image classification, uncertainty quantification, and learning efficiency. We compare these results with fully stochastic, partially stochastic and deterministic baselines, and we perform ablation studies to understand the impact of key hyperparameters.
Researcher Affiliation Academia 1Oxford-Man Institute of Quantitative Finance, University of Oxford 2Mathematical Institute, University of Oxford 3Department of Mathematics, Imperial College London.
Pseudocode No The paper describes methods and processes through prose and mathematical formulations, but it does not include any clearly labeled pseudocode or algorithm blocks.
Open Source Code Yes The code can be found at https://github.com/Sergio20f/part_stoch_inf_deep
Open Datasets Yes For the MNIST (Le Cun et al., 2010) dataset, a single SDE-BNN block is utilised, whereas, for CIFAR-10 (Krizhevsky, 2009), we employ a multi-scale approach with several SDE-BNN blocks with the invertible downsampling from Dinh et al. (2016) in between.
Dataset Splits No The paper references the use of training, validation, and test accuracies/ECEs in figures and discussions, but it does not explicitly state the specific dataset split percentages or sample counts for these divisions.
Hardware Specification Yes Lastly, experiments were conducted on a single Nvidia RTX 3090 GPU.
Software Dependencies No The paper provides hyperparameter settings and mentions general activation functions and solvers, but it does not specify software dependencies with version numbers (e.g., Python, PyTorch, or CUDA versions).
Experiment Setup Yes Table 5. Hyperparameter settings for model evaluations in the classification tasks as presented in Table 1. Experiments were conducted on a single Nvidia RTX 3090 GPU within our compute clusters. Throughout the training, no scheduling was applied to the hyperparameters.