Notice: The reproducibility variables underlying each score are classified using an automated LLM-based pipeline, validated against a manually labeled dataset. LLM-based classification introduces uncertainty and potential bias; scores should be interpreted as estimates. Full accuracy metrics and methodology are described in [1].

Taxonomy-Structured Domain Adaptation

Authors: Tianyi Liu, Zihao Xu, Hao He, Guang-Yuan Hao, Guang-He Lee, Hao Wang

ICML 2023 | Venue PDF | LLM Run Details

Reproducibility Variable Result LLM Response
Research Type Experimental Empirically, our method achieves state-of-the-art performance on both synthetic and real-world datasets with successful adaptation.
Researcher Affiliation Academia 1Rutgers University 2Massachusetts Institute of Technology 3The Chinese University of Hong Kong.
Pseudocode Yes We summarize the training procedure formally in Algorithm 1. ... The inference procedure of TSDA is formally presented in Algorithm 2.
Open Source Code Yes Code is available at https: //github.com/Wang-ML-Lab/TSDA.
Open Datasets Yes Image Net-Attribute-DT (Ouyang et al., 2015) builds on the animal images from Image Net... CUB-DT (He & Peng, 2019) contains 11,788 images of 200 bird categories.
Dataset Splits No The paper describes source and target domains for training and testing, but it does not specify explicit validation dataset splits (e.g., percentages, counts, or predefined sets) for model tuning or evaluation beyond the source-target distinction.
Hardware Specification Yes All experiments are run on NVDIA Ge Force RTX 2080 Ti GPUs.
Software Dependencies No The paper mentions using 'Py Torch' and 'Adam optimizer (Kingma & Ba, 2015)' but does not provide specific version numbers for these or other software dependencies.
Experiment Setup Yes We have λd, λt and λe as the weights that balance the discriminator loss, the taxonomist loss, and the predictor loss. ... λd and λe range from 0.1 to 1 and λt ranges from 0.1 to 10. ... We use Adam optimizer (Kingma & Ba, 2015) for all models with learning rates from 1 10 4 to 1 10 6. ... Table 4 shows the experiment results of various λd, λt combination on DT-14.