Latent Bernoulli Autoencoder

Authors: Jiri Fajtl, Vasileios Argyriou, Dorothy Monekosso, Paolo Remagnino

ICML 2020 | Conference PDF | Archive PDF | Plain Text | LLM Run Details

Reproducibility Variable Result LLM Response
Research Type Experimental Our method performs on a par or better than the current state-of-the-art methods on common Celeb A, CIFAR-10 and MNIST datasets.
Researcher Affiliation Academia 1Kingston University, London, UK 2Leeds Beckett University, Leeds, UK.
Pseudocode Yes Algorithm 1 Latent s to hyperplane normal r inversion
Open Source Code Yes Py Torch code and trained models are publicly available on github1. 1https://github.com/ok1zjf/lbae
Open Datasets Yes We trained and tested our model on the Celeb A (Liu et al., 2015), CIFAR10 (Krizhevsky & Hinton, 2009) and MNIST (Le Cun et al., 2010) datasets
Dataset Splits No with the default train/test splits
Hardware Specification No The paper does not provide specific details on the hardware used for running experiments, such as GPU or CPU models.
Software Dependencies No The paper mentions 'Py Torch code' but does not specify a version number for PyTorch or any other software dependencies.
Experiment Setup Yes The model was trained with ADAM(Kingma & Ba, 2015) with learning rate 10 3, no weight decay and 512 batch size. Mean squared error is used as the reconstruction loss except for MNIST where we use the binary cross entropy. Table 3 specifies epochs: MNIST 2000, CIFAR-10 2000, CELEBA 500.