Notice: The reproducibility variables underlying each score are classified using an automated LLM-based pipeline, validated against a manually labeled dataset. LLM-based classification introduces uncertainty and potential bias; scores should be interpreted as estimates. Full accuracy metrics and methodology are described in [1].

Lipschitz Lifelong Reinforcement Learning

Authors: Erwan Lecarpentier, David Abel, Kavosh Asadi, Yuu Jinnai, Emmanuel Rachelson, Michael L. Littman8270-8278

AAAI 2021 | Venue PDF | LLM Run Details

Reproducibility Variable Result LLM Response
Research Type Experimental We illustrate the benefits of the method in Lifelong RL experiments. ... We illustrate the benefits of the method in Lifelong RL experiments. ... Section 5: Experiments
Researcher Affiliation Collaboration 1ISAE-SUPAERO, Universit e de Toulouse, France 2ONERA, The French Aerospace Lab, Toulouse, France 3Brown University, Providence, Rhode Island, USA 4Amazon Web Service, Palo Alto, California, USA
Pseudocode Yes Algorithm 1: Lipschitz RMax algorithm
Open Source Code Yes Code available at https://github.com/Su Re LI/llrl
Open Datasets Yes The environment used in all experiments is a variant of the tight task used by Abel et al. (2018).
Dataset Splits No The paper describes the experimental setup in an RL context, including sampling tasks sequentially and running episodes. It does not provide explicit training/validation/test dataset splits as typically found in supervised learning, nor does it specify cross-validation folds. It mentions repeating operations '10 times to narrow the confidence intervals' which refers to overall experimental runs rather than dataset splitting.
Hardware Specification No The paper does not provide specific details about the hardware used to run the experiments, such as CPU or GPU models, memory specifications, or cloud computing instance types.
Software Dependencies No The paper does not list specific software dependencies or libraries with their version numbers (e.g., Python 3.8, PyTorch 1.9, TensorFlow 2.x) that are required to reproduce the experiments.
Experiment Setup Yes We set nknown = 10, δ = 0.05, and ϵ = 0.01 (discussion in Appendix, Section 15). ... We sample 15 tasks sequentially among this set, each run for 2000 episodes of length 10.