Notice: The reproducibility variables underlying each score are classified using an automated LLM-based pipeline, validated against a manually labeled dataset. LLM-based classification introduces uncertainty and potential bias; scores should be interpreted as estimates. Full accuracy metrics and methodology are described in [1].
Likelihood Training of Schrödinger Bridge using Forward-Backward SDEs Theory
Authors: Tianrong Chen, Guan-Horng Liu, Evangelos Theodorou
ICLR 2022 | Venue PDF | LLM Run Details
| Reproducibility Variable | Result | LLM Response |
|---|---|---|
| Research Type | Experimental | we show that the resulting training algorithm achieves comparable results on generating realistic images on MNIST, Celeb A, and CIFAR10. Our code is available at https://github.com/ghliu/SB-FBSDE. |
| Researcher Affiliation | Academia | Tianrong Chen , Guan-Horng Liu , Evangelos A. Theodorou Georgia Institute of Technology, USA EMAIL |
| Pseudocode | Yes | Algorithm 1 Likelihood training of SB-FBSDE, Algorithm 2 Joint (diffusion flow-based) training, Algorithm 3 Alternate (IPF-based) training, Algorithm 4 Generative Process of SB-FBSDE |
| Open Source Code | Yes | Our code is available at https://github.com/ghliu/SB-FBSDE. |
| Open Datasets | Yes | We testify SB-FBSDE on two toy datasets and three image datasets, i.e. MNIST, Celeb A, and CIFAR10. |
| Dataset Splits | No | No explicit statements on training, validation, or test dataset splits (e.g., percentages or counts) are provided for MNIST, Celeb A, or CIFAR10. While these are standard datasets with common splits, the paper does not specify them. |
| Hardware Specification | Yes | All networks adopt position encoding and are trained with Adam W (Loshchilov & Hutter, 2017) on a TITAN RTX. |
| Software Dependencies | No | The paper does not provide specific software dependencies with version numbers (e.g., Python, PyTorch, or TensorFlow versions) that are required to replicate the experiment. |
| Experiment Setup | Yes | We testify SB-FBSDE on two toy datasets and three image datasets, i.e. MNIST, Celeb A, and CIFAR10. pprior is set to a zero-mean Gaussian whose variance varies for each task and can be computed according to Song & Ermon (2020). We parameterize Z( , ; θ) and b Z( , ; φ) with residual-based networks for toy datasets and consider Unet (Ronneberger et al., 2015) and NCSN++ (Song et al., 2020) respectively for MNIST/Celeb A and CIFAR10. All networks adopt position encoding and are trained with Adam W (Loshchilov & Hutter, 2017) on a TITAN RTX. We adopt VE-SDE (i.e. f := 0; see Song et al. (2020)) as our SDE backbone... On all datasets, we set the horizon T =1.0 and solve the SDEs via the Euler-Maruyama method. The interval [0, T] is discretized into 200 steps for CIFAR10 and 100 steps for all other datasets... Other details are left in Appendix D. |