CONTRAST: Continual Multi-source Adaptation to Dynamic Distributions
Authors: Sk Miraj Ahmed, Fahim Faisal Niloy, Xiangyu Chang, Dripta S. Raychaudhuri, Samet Oymak, Amit Roy-Chowdhury
NeurIPS 2024 | Conference PDF | Archive PDF | Plain Text | LLM Run Details
| Reproducibility Variable | Result | LLM Response |
|---|---|---|
| Research Type | Experimental | Experimental analysis on diverse datasets demonstrates that the combination of multiple source models does at least as well as the best source (with hindsight knowledge), and performance does not degrade as the test data distribution changes over time (robust to forgetting). |
| Researcher Affiliation | Collaboration | 1University of California, Riverside, 2Brookhaven National Laboratory, 3AWS AI Labs, 4University of Michigan, Ann Arbor |
| Pseudocode | Yes | Algorithm 1: Overview of CONTRAST |
| Open Source Code | No | Question: Does the paper provide open access to the data and code, with sufficient instructions to faithfully reproduce the main experimental results, as described in supplemental material? Answer: [No] |
| Open Datasets | Yes | Datasets. We demonstrate the efficacy of our approach using both static target distribution and dynamic target data distributions. For static case, we employ the Digits and Office-Home datasets [42]. For the dynamic case, we utilize CIFAR-100C and CIFAR-10C [43]. |
| Dataset Splits | No | No explicit training/validation/test splits are provided for the datasets used in the experiments with specific percentages or counts. The theoretical section mentions 'validation(ν) distribution D' but this does not specify the splits used in the empirical evaluations. |
| Hardware Specification | No | No specific hardware details (e.g., GPU/CPU models, memory amounts) are provided for running the experiments. The NeurIPS checklist mentions 'GPU' generally but no specific models or configurations. |
| Software Dependencies | No | The paper mentions software components like 'Adam optimizer' and uses frameworks implicitly (e.g., ResNet-18 models), but it does not provide specific version numbers for programming languages, libraries, or frameworks (e.g., Python, PyTorch, CUDA versions). |
| Experiment Setup | Yes | We use Res Net-18 [44] model for all our experiments. For solving the optimization of Eq. (4), we first initialize the combination weights using Eq. (6) and calculate the optimal learning rate using Eq. (7). After that, we use 5 iterations to update the combination weights using SGD optimizer and the optimal learning rate. For all the experiments we use a batch size of 128, as used by Tent [3]. For more details on implementation and experimental setting see Appendix. |