Boosting Single Positive Multi-label Classification with Generalized Robust Loss

Authors: Yanxi Chen, Chunxiao Li, Xinyang Dai, Jinhuan Li, Weiyu Sun, Yiming Wang, Renyuan Zhang, Tinghe Zhang, Bo Wang

IJCAI 2024 | Conference PDF | Archive PDF | Plain Text | LLM Run Details

Reproducibility Variable Result LLM Response
Research Type Experimental Extensive experiments show that our approach can significantly improve SPML performance and outperform the vast majority of state-of-the-art methods on all the four benchmarks. Our code is available at https://github.com/yan4xi1/GRLoss. Our contributions can be summarized in four-fold: Experimental level: We demonstrate the superiority of our GR Loss by conducting extensive empirical analysis with performance comparison against state-of-the-art SPML methods across four benchmarks.
Researcher Affiliation Academia Yanxi Chen , Chunxiao Li , Xinyang Dai , Jinhuan Li , Weiyu Sun , Yiming Wang , Renyuan Zhang , Tinghe Zhang and Bo Wang University of International Business and Economics {202310411355,202310411351,wangbo}@uibe.edu.cn
Pseudocode No The paper provides mathematical formulations and descriptions of functions, but it does not include a dedicated pseudocode block or algorithm listing.
Open Source Code Yes Our code is available at https://github.com/yan4xi1/GRLoss.
Open Datasets Yes We evaluate our proposed GR Loss on four benchmark datasets: Pascal VOC-2012 (VOC) [Everingham and Winn, 2012], MS-COCO-2014(COCO) [Lin et al., 2014], NUS-WIDE(NUS) [Chua et al., 2009], and CUB200-2011(CUB) [Wah et al., 2011].
Dataset Splits Yes We first simulate the single-positive label training environments commonly used in SPML [Cole et al., 2021], and replicate their training, validation and testing samples. In these datasets, only one positive label is randomly selected for each training instance, while the validation and test sets remain fully labeled.
Hardware Specification No The paper mentions employing 'Res Net-50 architecture' and being 'pre-trained on Image Net dataset' but does not specify the hardware (e.g., GPU, CPU models, memory) used to run the experiments.
Software Dependencies No The paper mentions using a 'Res Net-50 architecture' and refers to hyperparameter settings, but it does not provide specific version numbers for software dependencies such as Python, PyTorch, or CUDA.
Experiment Setup Yes Each image is resized into 448 × 448, and performed data augmentation by randomly flipping an image horizontally. We initially conduct a search to determine and fix the hyperparameters q2 and q3 in Eq.(14), typically 0.01 and 1, respectively. ... Therefore, we only need to adjust four hyperparameters in (β(T ), α(T )). More details about hyperparameter settings are described in Appendix E.2 of our full version.