Notice: The reproducibility variables underlying each score are classified using an automated LLM-based pipeline, validated against a manually labeled dataset. LLM-based classification introduces uncertainty and potential bias; scores should be interpreted as estimates. Full accuracy metrics and methodology are described in [1].

Lifted Weighted Mini-Bucket

Authors: Nicholas Gallo, Alexander T. Ihler

NeurIPS 2018 | Venue PDF | LLM Run Details

Reproducibility Variable Result LLM Response
Research Type Experimental Experimental results demonstrate the utility of this class of approximations, especially in models with strong repulsive potentials.
Researcher Affiliation Academia Nicholas Gallo University of California Irvine Irvine, CA 92637-3435 EMAIL Alexander Ihler University of California Irvine Irvine, CA 92637-3435 EMAIL
Pseudocode Yes Algorithm 1 summarizes the LWMB tree construction algorithm (similar to ground mini-bucket construction [7]) developped in this section.
Open Source Code No The paper does not provide any explicit statement about open-sourcing the code for the described methodology, nor does it include a link to a code repository.
Open Datasets No The paper describes how the data for the experiments was generated synthetically ('We run experiments with N = | | = 512, with clustered evidence. We randomly assign elements of to one of K = 16 clusters...'), but it does not refer to a publicly available dataset or provide access information.
Dataset Splits No The paper does not provide specific train/validation/test dataset splits, percentages, or sample counts needed to reproduce the data partitioning. It only describes the synthetic generation process for the data used.
Hardware Specification No The paper does not provide any specific hardware details such as GPU models, CPU models, or memory specifications used for running the experiments.
Software Dependencies No The paper mentions that 'code has been written in C++', but it does not specify any software dependencies, libraries, or solvers with version numbers required to replicate the experiments.
Experiment Setup Yes We run experiments with N = | | = 512, with clustered evidence... We randomly assign elements of to one of K = 16 clusters... Each cluster generates a (scalar) center on N(0, 2) each member of the cluster is then perturbed from its center by N(0, 0.4) noise... We call a black-box convex optimization (using non-linear conjugate gradients) allowing a maximum of 1000 function evaluations.