Notice: The reproducibility variables underlying each score are classified using an automated LLM-based pipeline, validated against a manually labeled dataset. LLM-based classification introduces uncertainty and potential bias; scores should be interpreted as estimates. Full accuracy metrics and methodology are described in [1].

Can contrastive learning avoid shortcut solutions?

Authors: Joshua Robinson, Li Sun, Ke Yu, Kayhan Batmanghelich, Stefanie Jegelka, Suvrit Sra

NeurIPS 2021 | Venue PDF | LLM Run Details

Reproducibility Variable Result LLM Response
Research Type Experimental Empirically, we observe that IFM reduces feature suppression, and as a result improves performance on vision and medical imaging tasks. We train encoders with Res Net-18 backbone using Sim CLR [5]. To study correlations between the loss value and error on downstream tasks, we train 33 encoders on Trifeature and 7 encoders on STL-digits with different hyperparameter settings (see App. C.2 for full details on training and hyperparameters).
Researcher Affiliation Academia Joshua Robinson MIT CSAIL & LIDS EMAIL Li Sun University of Pittsburgh EMAIL Ke Yu University of Pittsburgh EMAIL Kayhan Batmanghelich University of Pittsburgh EMAIL Stefanie Jegelka MIT CSAIL EMAIL MIT LIDS EMAIL
Pseudocode No The paper describes methods using mathematical formulations and descriptive text, but does not include a clearly labeled 'Pseudocode' or 'Algorithm' block.
Open Source Code Yes The code is available at: https://github. com/joshr17/IFM.
Open Datasets Yes We use two datasets with known semantic features: (1) In the Trifeature data, [16] each image is 128 × 128 and has three features: color, shape, and texture... and benchmarks IFM on Image Net100 [44] using Mo Co-v2... and COPDGene dataset [38]
Dataset Splits Yes All encoders are evaluated using the test accuracy of a linear classifier trained on the full training dataset (see Appdx. C.4 for full setup details). The values are the average of 5-fold cross validation with standard deviations.
Hardware Specification No The paper does not provide specific hardware details such as GPU models, CPU types, or memory specifications used for running the experiments.
Software Dependencies No The paper mentions software like PyTorch and scikit-learn in its references but does not provide specific version numbers for these or any other software dependencies required for replication.
Experiment Setup Yes We train encoders with Res Net-18 backbone using Sim CLR [5]. All encoders have Res Net-50 backbones and are trained for 400 epochs (with the exception of on Image Net100, which is trained for 200 epochs). we train Res Net-18 encoders for 200 epochs with τ ∈ {0.05, 0.2, 0.5} and IFM using ε = 0.1 for simplicity. (see Appdx. C.4.1 for full setup details).