Notice: The reproducibility variables underlying each score are classified using an automated LLM-based pipeline, validated against a manually labeled dataset. LLM-based classification introduces uncertainty and potential bias; scores should be interpreted as estimates. Full accuracy metrics and methodology are described in [1].

Long-Tailed Out-of-Distribution Detection via Normalized Outlier Distribution Adaptation

Authors: Wenjun Miao, Guansong Pang, Jin Zheng, Xiao Bai

NeurIPS 2024 | Venue PDF | LLM Run Details

Reproducibility Variable Result LLM Response
Research Type Experimental Empirical results on three popular benchmarks for OOD detection in LTR show the superior performance of Adapt OD over state-of-the-art methods.
Researcher Affiliation Academia Wenjun Miao Beihang University, China EMAIL; Guansong Pang Singapore Management University, Singapore EMAIL; Jin Zheng Beihang University, China EMAIL; Xiao Bai Beihang University, China EMAIL
Pseudocode Yes Algorithm 1 : Adapt OD Training
Open Source Code Yes Code is available at https://github.com/mala-lab/Adapt OD.
Open Datasets Yes Following [30, 40, 45], we use three popular long-tailed datasets CIFAR10-LT [3], CIFAR100-LT [3] and Image Net-LT [26] as ID data Xin. The default imbalance ratio is set to ρ = 100 on CIFAR10/100-LT. Tiny Images80M [38] is used as the outlier data Xaux_out for CIFAR10/100-LT and Image Net-Extra [40] is used as outlier data for Image Net-LT.
Dataset Splits Yes For ID datasets, the original version of CIFAR10 [10] and CIFAR100 [10] contains 50, 000 training images and 10, 000 validation images of size 32 32 with 10 and 100 classes, respectively. CIFAR10LT and CIFAR100-LT are the imbalanced version of them, which reduce the number of training examples per class and keep the validation set unchanged.
Hardware Specification Yes All experiences are performed with 8 NVIDIA RTX 3090.
Software Dependencies No The paper mentions 'SGD optimizer' and 'cosine annealing learning rate scheduler' but does not specify software versions for libraries like PyTorch, TensorFlow, or specific Python versions within the main text or appendices for reproducibility.
Experiment Setup Yes For experiments on CIFAR10-LT [3] and CIFAR100-LT [3], we pre-train our model based on Res Net18 [11] for 320 epochs with an initial learning rate 0.01 [1, 4] using only cross-entropy loss and fine-tune the linear classifier of this model for 20 epochs with an initial learning rate 0.001 [4, 24]. The batch size is 64 for ID data at the pre-training stage, 128 for ID data at the fine-tuning stage, and 256 for outlier data at the fine-tuning stage [4,30,40].