Towards Enabling Meta-Learning from Target Models

Authors: Su Lu, Han-Jia Ye, Le Gan, De-Chuan Zhan

NeurIPS 2021 | Conference PDF | Archive PDF | Plain Text | LLM Run Details

Reproducibility Variable Result LLM Response
Research Type Experimental We empirically verify the effectiveness of S/T protocol in a typical application of meta-learning, i.e., few-shot learning.
Researcher Affiliation Academia Su Lu Han-Jia Ye Le Gan De-Chuan Zhan State Key Laboratory for Novel Software Technology Nanjing University, Nanjing, 210023, China {lus,yehj}@lamda.nju.edu.cn, {ganl,zhandc}@nju.edu.cn
Pseudocode No The paper does not contain any clearly labeled pseudocode or algorithm blocks.
Open Source Code Yes Our code is available at https://github.com/njulus/ST.
Open Datasets Yes In this part, we evaluate our S/T protocol on two benchmark datasets, i.e., mini Image Net [24] and tiered Image Net [17].
Dataset Splits Yes 10000 tasks are used for both meta-training and meta-testing. 500 tasks are used for meta-validation. There are 64 classes for meta-training, 16 classes for meta-validation, and 20 classes for meta-testing.
Hardware Specification Yes We run the experiment on an Nvidia Ge Force RTX 2080ti GPU and Intel(R) Xeon(R) Silver 4110 CPU.
Software Dependencies No The paper mentions models like Res Net-12 and common frameworks are implied by the algorithms used (e.g., MAML, Proto Net), but it does not specify explicit software dependencies with version numbers (e.g., PyTorch 1.9, Python 3.8).
Experiment Setup Yes For each task, we generate 10 support instances by uniformly sampling x in range [ 5, 5]. For S/Q protocol, we additionally sample 30 query instances for each task. We set λ to 0.8, a relatively large value, in most of experiments. We pre-train a Res Net-12 with a linear layer on the meta-training split of mini Image Net.