Notice: The reproducibility variables underlying each score are classified using an automated LLM-based pipeline, validated against a manually labeled dataset. LLM-based classification introduces uncertainty and potential bias; scores should be interpreted as estimates. Full accuracy metrics and methodology are described in [1].

Label Delay in Online Continual Learning

Authors: Botos Csaba, Wenxuan Zhang, Matthias Mรผller, Ser Nam Lim, Philip Torr, Adel Bibi

NeurIPS 2024 | Venue PDF | LLM Run Details

Reproducibility Variable Result LLM Response
Research Type Experimental In our extensive experiments amounting to 25000 GPU hours, we show that merely increasing the computational resources is insufficient to tackle this challenge.
Researcher Affiliation Collaboration 1University of Oxford 2King Abdullah University of Science and Technology 3Intel Labs 4University of Central Florida
Pseudocode Yes Algorithm 1 Single OCL time step with Label Delay
Open Source Code Yes The implementation for reproducing our experiments can be found at https://github.com/botcs/label-delay-exp.
Open Datasets Yes We conduct our experiments on four large-scale online continual learning datasets, Continual Localization (CLOC) [4], Continual Google Landmarks (CGLM) [5], Functional Map of the World (FMo W) [6], and Yearbook [7].
Dataset Splits Yes We follow the same training and validation set split of CLOC as in [4] and o CGLM as in [5] and the official released splits for FMo W [6] and Yearbook [7].
Hardware Specification Yes Most of the experiments are using a single A100 GPU with 12 CPU.
Software Dependencies No The paper does not provide specific ancillary software details with version numbers (e.g., library or solver names with version numbers like Python 3.8, CPLEX 12.4) needed to replicate the experiment.
Experiment Setup Yes We use SGD with the learning rate of 0.005, momentum of 0.9, and weight decay of 10โˆ’5. We apply random cropping and resizing to the images, such that the resulting input has a resolution of 224 ร— 224.