Notice: The reproducibility variables underlying each score are classified using an automated LLM-based pipeline, validated against a manually labeled dataset. LLM-based classification introduces uncertainty and potential bias; scores should be interpreted as estimates. Full accuracy metrics and methodology are described in [1].

Faster Discrete Convex Function Minimization with Predictions: The M-Convex Case

Authors: Taihei Oki, Shinsaku Sakaue

NeurIPS 2023 | Venue PDF | LLM Run Details

Reproducibility Variable Result LLM Response
Research Type Experimental Experiments in Section 5 confirm that using predictions can improve empirical computation costs.
Researcher Affiliation Academia Taihei Oki The University of Tokyo Tokyo, Japan EMAIL Shinsaku Sakaue The University of Tokyo Tokyo, Japan EMAIL
Pseudocode Yes Algorithm 1 Greedy algorithm
Open Source Code Yes The source code is available at https://github.com/ssakaue/alps-m-convex-code.
Open Datasets No We thus create a dataset of T = 100 random instances for each σ {1, 5, 10, 20}. We generate 10 such random datasets independently to calculate the mean and standard deviation of the results. We then generate T = 100 instances by perturbing parameters defining constraints and objectives with Gaussian noises multiplied by σ = 0.1, 1.0, 10.0
Dataset Splits No The paper describes generating instances and using online learning, but it does not specify explicit training, validation, and test splits for a static dataset.
Hardware Specification Yes We used Mac Book Air with Apple M2 chip, 24 GB of memory, and mac OS Ventura 13.2.1.
Software Dependencies Yes We implemented algorithms in Python 3.9.12 with libraries such as Num Py 1.23.2. We used Gurobi 10.0.1 [17]
Experiment Setup Yes We set c Y = max{P i Y i + σua, 1}, where ua follows the standard normal distribution and σ controls the noise strength. We let ℓY = min{2h + ub, R/2n h}, where h {0, 1, . . . , log n} is the height of Y in TF (a leaf Y has h = 0) and ub is drawn uniformly at random from {0, 1, . . . , 50}; the minimum with R/2n h is taken to ensure that the feasible region is non-empty. ... We learn predictions ˆxt RN for t = 1, . . . , T by using OSD on V = { x [0, R]N : x(N) = R } with a step size of 0.01 pR/n