Continual Learning of a Mixed Sequence of Similar and Dissimilar Tasks

Authors: Zixuan Ke, Bing Liu, Xingchang Huang

NeurIPS 2020 | Conference PDF | Archive PDF | Plain Text | LLM Run Details

Reproducibility Variable Result LLM Response
Research Type Experimental Empirical evaluation using sequences of mixed tasks demonstrates the effectiveness of the proposed model.
Researcher Affiliation Academia Zixuan Ke1, Bing Liu1, and Xingchang Huang2 1 Department of Computer Science, University of Illinois at Chicago 2 ETH Zurich {zke4, liub}@uic.edu, huangxch3@gmail.com
Pseudocode No The paper describes the model and methods using text and mathematical equations (e.g., Equations 1-11) and diagrams (Figure 1), but does not include any explicit pseudocode or algorithm blocks.
Open Source Code Yes 2https://github.com/Zixuan Ke/CAT
Open Datasets Yes We adopt two similar-task datasets from federated learning... from two publicly available federated learning datasets (Caldas et al., 2018)... EMNIST (Le Cun et al., 1998) and CIFAR100 (Krizhevsky et al., 2009).
Dataset Splits Yes We further split about 10% of the original training data as the validate data.
Hardware Specification No The paper describes network architectures (e.g., '2-layer fully connected network', 'CNN based Alex Net-like architecture') and training details, but it does not specify any particular hardware components like GPU or CPU models used for the experiments.
Software Dependencies No The paper mentions using SGD for training and describes network architectures, but it does not provide specific version numbers for software dependencies like programming languages, libraries, or frameworks (e.g., Python, PyTorch, TensorFlow, CUDA).
Experiment Setup Yes We use 140 for smax in s, dropout of 0.5 between fully connected layers... We set the number of attention heads to 5... We train all models using SGD with the learning rate of 0.05... The batch size is set to 64.