Notice: The reproducibility variables underlying each score are classified using an automated LLM-based pipeline, validated against a manually labeled dataset. LLM-based classification introduces uncertainty and potential bias; scores should be interpreted as estimates. Full accuracy metrics and methodology are described in [1].

Continuous Regularized Wasserstein Barycenters

Authors: Lingxiao Li, Aude Genevay, Mikhail Yurochkin, Justin M. Solomon

NeurIPS 2020 | Venue PDF | LLM Run Details

Reproducibility Variable Result LLM Response
Research Type Experimental We demonstrate the effectiveness of our approach and compare against previous work on synthetic examples and real-world applications.
Researcher Affiliation Collaboration Lingxiao Li MIT CSAIL EMAIL Aude Genevay MIT CSAIL EMAIL Mikhail Yurochkin IBM Research, MIT-IBM Watson AI Lab EMAIL Justin Solomon MIT CSAIL, MIT-IBM Watson AI Lab EMAIL
Pseudocode Yes Algorithm 1: Stochastic gradient descent to solve the regularized barycenter problem (11)
Open Source Code Yes The source code is publicly available at https://github.com/lingxiaoli94/CWB.
Open Datasets Yes We consider Poisson regression for the task of predicting the hourly number of bike rentals using features such as the day of the week and weather conditions.3 (Footnote 3: http://archive.ics.uci.edu/ml/datasets/Bike+Sharing+Dataset)
Dataset Splits No The paper describes splitting the data into 5 equally-sized subsets for posterior aggregation, but does not explicitly mention a 'validation set' or a dedicated validation split in the context of model training or evaluation.
Hardware Specification Yes We ran our experiments using a NVIDIA Tesla V100 GPU on a Google cloud instance with 12 computeoptimized CPUs and 64GB memory.
Software Dependencies Yes The stochastic gradient descent used to solve (11) and (14) is implemented in Tensor๏ฌ‚ow 2.1 [Aba+16].
Experiment Setup Yes In all experiments below, we use Adam optimizer [KB14] with learning rate 10 4 and batch size 4096 or 8192 for the training. The dual potentials {fi, gi}n i=1 in (11) are each parameterized as neural networks with two fully-connected layers (d ! 128 ! 256 ! 1) using Re LU activations. Every Ti in (14) is parameterized with layers (d ! 128 ! 256 ! d).