Notice: The reproducibility variables underlying each score are classified using an automated LLM-based pipeline, validated against a manually labeled dataset. LLM-based classification introduces uncertainty and potential bias; scores should be interpreted as estimates. Full accuracy metrics and methodology are described in [1].
Stochastic Optimal Control Matching
Authors: Carles Domingo i Enrich, Jiequn Han, Brandon Amos, Joan Bruna, Ricky T. Q. Chen
NeurIPS 2024 | Venue PDF | LLM Run Details
| Reproducibility Variable | Result | LLM Response |
|---|---|---|
| Research Type | Experimental | Experimentally, our algorithm achieves lower error than all the existing IDO techniques for stochastic optimal control for three out of four control problems, in some cases by an order of magnitude. |
| Researcher Affiliation | Collaboration | Carles Domingo-Enrich NYU & FAIR, Meta EMAIL Jiequn Han Flatiron Institute EMAIL Brandon Amos FAIR, Meta EMAIL Joan Bruna NYU & Flatiron Institute EMAIL Ricky T. Q. Chen FAIR, Meta EMAIL |
| Pseudocode | Yes | Algorithm 1 Iterative Diffusion Optimization (IDO) algorithms for stochastic optimal control |
| Open Source Code | Yes | Code can be found at https://github.com/facebookresearch/SOC-matching. |
| Open Datasets | No | The paper defines the functions and parameters for its experimental settings (QUADRATIC ORNSTEIN UHLENBECK, LINEAR ORNSTEIN UHLENBECK, DOUBLE WELL, and PATH INTEGRAL SAMPLER ON MIXTURE OF GAUSSIANS) rather than using pre-existing, publicly available datasets. |
| Dataset Splits | No | The paper describes dynamic simulation of trajectories and Monte Carlo estimation for evaluation, specifying batch sizes for these simulations, but does not refer to static training/validation/test dataset splits in the conventional sense for a fixed dataset. |
| Hardware Specification | Yes | Each algorithm was run using a 16GB V100 GPU. |
| Software Dependencies | No | The paper mentions using 'Adam' for optimization but does not provide specific version numbers for software libraries, programming languages (e.g., Python, PyTorch), or other dependencies. |
| Experiment Setup | Yes | For all them, we train the control using Adam with learning rate 1 10 4. For SOCM, we train the reparametrization matrices using Adam with learning rate 1 10 2. We use batch size m = 128 unless otherwise specified. |