Sampling with Mollified Interaction Energy Descent

Authors: Lingxiao Li, qiang liu, Anna Korba, Mikhail Yurochkin, Justin Solomon

ICLR 2023 | Conference PDF | Archive PDF | Plain Text | LLM Run Details

Reproducibility Variable Result LLM Response
Research Type Experimental We show experimentally that for unconstrained sampling problems, our algorithm performs on par with existing particle-based algorithms like SVGD, while for constrained sampling problems our method readily incorporates constrained optimization techniques to handle more flexible constraints with strong performance compared to alternatives.
Researcher Affiliation Collaboration Lingxiao Li MIT CSAIL lingxiao@mit.edu Qiang Liu University of Texas at Austin lqiang@cs.utexas.edu Anna Korba CREST, ENSAE, IP Paris anna.korba@ensae.fr Mikhail Yurochkin IBM Research, MIT-IBM Watson AI Lab mikhail.yurochkin@ibm.com Justin Solomon MIT CSAIL jsolomon@mit.edu
Pseudocode Yes Algorithm 1: Mollified interaction energy descent (MIED) in the logarithmic domain.
Open Source Code Yes The source code can be found at https://github.com/lingxiaoli94/MIED.
Open Datasets Yes Fairness Bayesian neural networks. We train fair Bayesian neural networks to predict whether the annual income of a person is at least $50, 000 with gender as the protected attribute using the Adult Income dataset (Kohavi et al., 1996).
Dataset Splits Yes We use 80%/20% training/test split.
Hardware Specification No The paper does not provide specific hardware details (e.g., GPU/CPU models, memory amounts) used for running its experiments.
Software Dependencies No All methods by default use a learning rate of 0.01 with Adam optimizer (Kingma & Ba, 2014). The paper mentions the Adam optimizer but does not specify its version or the versions of any other software libraries or programming languages used.
Experiment Setup Yes All methods by default use a learning rate of 0.01 with Adam optimizer (Kingma & Ba, 2014). All methods are run with identical initialization and learning rate 0.01. Results are reported after 10^4 iterations.