GFlowNets and variational inference
Authors: Nikolay Malkin, Salem Lahlou, Tristan Deleu, Xu Ji, Edward J Hu, Katie E Everett, Dinghuai Zhang, Yoshua Bengio
ICLR 2023 | Conference PDF | Archive PDF | Plain Text | LLM Run Details
| Reproducibility Variable | Result | LLM Response |
|---|---|---|
| Research Type | Experimental | We demonstrate that, in certain cases, VI algorithms are equivalent to special cases of GFlow Nets in the sense of equality of expected gradients of their learning objectives. We then point out the differences between the two families and show how these differences emerge experimentally. |
| Researcher Affiliation | Collaboration | Nikolay Malkin , Salem Lahlou , Tristan Deleu Mila, Universit e de Montr eal Xu Ji, Edward Hu Mila, Universit e de Montr eal Katie Everett Google Research Dinghuai Zhang Mila, Universit e de Montr eal Yoshua Bengio Mila, Universit e de Montr eal, CIFAR |
| Pseudocode | No | The paper does not contain structured pseudocode or algorithm blocks. Appendix A contains a figure illustrating a concept, not an algorithm. |
| Open Source Code | Yes | Code: https://github.com/GFNOrg/GFN_vs_HVI. |
| Open Datasets | No | The paper mentions using a 'synthetic hypergrid environment introduced by Bengio et al. (2021a)' and that 'for each experiment, we sampled a dataset D of 100 observations'. While these refer to data, no concrete access information (link, DOI, specific citation for the dataset itself) is provided for public access to the datasets used in the experiments. |
| Dataset Splits | No | The paper mentions 'a held-out set' for molecule experiments but does not specify exact split percentages, absolute sample counts, or reference predefined splits with citations for training, validation, or test sets across all experiments. |
| Hardware Specification | No | The paper states: 'This research was enabled in part by computational resources provided by the Digital Research Alliance of Canada.' This is a general statement and does not provide specific hardware details (e.g., CPU/GPU models, memory). |
| Software Dependencies | No | The paper mentions using 'Adam optimizer' and 'SGD' but does not provide specific version numbers for these software components or any other libraries/packages used. |
| Experiment Setup | Yes | The forward and backward policies are parametrized as neural networks with 2 hidden layers of 256 units each. We use a batch size of 64 for all learning objectives... The Adam optimizer... All models were trained with the Adam optimizer and batch size 4 for a maximum of 50000 batches. We used the Adam optimizer, with the best learning rate found among {10 6, 3 10 6, 10 5, 3 10 5, 10 4}. For the TB objective, we learned log 𝑍using SGD with a learning rate of 0.1 and momentum 0.8. |