Notice: The reproducibility variables underlying each score are classified using an automated LLM-based pipeline, validated against a manually labeled dataset. LLM-based classification introduces uncertainty and potential bias; scores should be interpreted as estimates. Full accuracy metrics and methodology are described in [1].

Federated Label-Noise Learning with Local Diversity Product Regularization

Authors: Xiaochen Zhou, Xudong Wang

AAAI 2024 | Venue PDF | LLM Run Details

Reproducibility Variable Result LLM Response
Research Type Experimental The results show that Fed LNL improves the test accuracy of a trained model by up to 25.98%, compared with the state-of-the-art FL schemes that tackle label-noise issues.The performance of Fed LNL is evaluated with extensive experiments. The effectiveness of the alternating update method and LDP regularizer is first verified in an ablation study. The overall performance of Fed LNL is then evaluated in the case of limited local training data samples.
Researcher Affiliation Academia Xiaochen Zhou, Xudong Wang Shanghai Jiao Tong University EMAIL, EMAIL
Pseudocode No The paper describes methods and processes verbally and with mathematical equations, but it does not include any formally labeled pseudocode or algorithm blocks.
Open Source Code No The paper does not contain any explicit statements about making its source code publicly available, nor does it provide a link to a code repository.
Open Datasets Yes Three datasets are adopted in the experiments: CIFAR-10 (Krizhevsky, Hinton et al. 2009), CIFAR-100 (Krizhevsky, Hinton et al. 2009), and Clothing1M (Xiao et al. 2015)).
Dataset Splits No The paper provides the total number of samples for training and testing datasets (e.g., 'CIFAR-10 ... # samples 50,000 ... # samples 10,000 (test)'), but it does not specify a separate validation split or its size.
Hardware Specification Yes All the experiments are executed on a server with one i9-10900k CPU, one Ge Force RTX 3090 GPU, and 64 GB RAM.
Software Dependencies Yes The implementation of Fed LNL is based on Pytorch-1.7.0 (Paszke et al. 2019).
Experiment Setup Yes In Fed LNL, each client trains its local classifier via stochastic gradient descent (SGD) with a momentum of 0.9. The learning rate is set to 0.01. The batch size is set to 64. The number of local iterations is set to 3 and the total number of communication rounds between the clients devices and the central server is set to 300. Hyperparameter λ is set to 0.01.