Notice: The reproducibility variables underlying each score are classified using an automated LLM-based pipeline, validated against a manually labeled dataset. LLM-based classification introduces uncertainty and potential bias; scores should be interpreted as estimates. Full accuracy metrics and methodology are described in [1].
Wasserstein Training of Restricted Boltzmann Machines
Authors: Grégoire Montavon, Klaus-Robert Müller, Marco Cuturi
NeurIPS 2016 | Venue PDF | LLM Run Details
| Reproducibility Variable | Result | LLM Response |
|---|---|---|
| Research Type | Experimental | We demonstrate their practical potential on data completion and denoising, for which the metric between observations plays a crucial role. |
| Researcher Affiliation | Academia | Grégoire Montavon Technische Universität Berlin EMAIL Klaus-Robert Müller Technische Universität Berlin EMAIL Marco Cuturi CREST, ENSAE, Université Paris-Saclay EMAIL Also with the Department of Brain and Cognitive Engineering, Korea University. |
| Pseudocode | No | The paper describes the steps of the Sinkhorn algorithm in paragraph form (e.g., 'repeat until u, v converge...'), but it does not present a formally labeled 'Pseudocode' or 'Algorithm' block. |
| Open Source Code | No | The paper does not contain any explicit statement about releasing source code or provide a link to a code repository. |
| Open Datasets | Yes | We consider three datasets: MNIST-small, a subsampled version of the original MNIST dataset [11] with only the digits 0 retained, a subset of the UCI PLANTS dataset [19] containing the geographical spread of plants species, and MNIST-code, 128-dimensional code vectors associated to each MNIST digit (additional details in the supplement). |
| Dataset Splits | Yes | We perform holdout validation on the quadratic containment coefficient η {10 4, 10 3, 10 2}, and on the KL weighting coefficient λ {0, 10 1, 100, 101, }. |
| Hardware Specification | No | The paper does not provide specific hardware details (e.g., CPU, GPU models, memory) used for running the experiments. |
| Software Dependencies | No | The paper does not specify any software dependencies with version numbers. |
| Experiment Setup | Yes | The number of hidden units of the RBM is set heuristically to 400 for all datasets. The learning rate is set heuristically to 0.01(λ 1) during the pretraining phase and modified to 0.01 min(1, λ 1) when training on the final objective. with smoothing parameter γ = 0.1 and distance D(x, x ) = H(x, x )/ H(x, x ) ˆp. |