Improved Modelling of Federated Datasets using Mixtures-of-Dirichlet-Multinomials

Authors: Jonathan Scott, Áine Cahill

ICML 2024 | Conference PDF | Archive PDF | Plain Text | LLM Run Details

Reproducibility Variable Result LLM Response
Research Type Experimental We demonstrate empirically that this algorithm is able to successfully infer meaningful parameters. We show that using these inferred parameters to create simulated clients on the server leads to more representative training simulations. Our experiments are implemented using the pfl-research framework (Granqvist et al., 2024).
Researcher Affiliation Collaboration 1Institute of Science and Technology Austria (ISTA) 2Apple. Correspondence to: Jonathan Scott <jonathan.scott@ist.ac.at>.
Pseudocode Yes Algorithm 1 Dirichlet-Multinomial Mixture Initialization Algorithm 2 Dirichlet-Multinomial Mixture MLE
Open Source Code Yes Our code can be found at https://github.com/apple/ pfl-research/tree/develop/publications/ mdm.
Open Datasets Yes We evaluate using synthetic data that follows the MDM distribution, CIFAR10 (Krizhevsky, 2009), FEMNIST (Caldas et al., 2018) and Folktables (Ding et al., 2021).
Dataset Splits Yes In Appendix A we outline a procedure for the server to choose the best value of K to use. ... 2. Sample a new cohort of clients that we have not yet seen and for each choice of K evaluate the log likelihood, quation 4, on this cohort of clients. 3. Use the K that gave the highest log likelihood on this validation cohort of clients.
Hardware Specification No The paper does not provide specific hardware details (e.g., GPU/CPU models, memory specifications) used for running its experiments. It mentions 'on device training' and 'mobile devices' in a general context but not for the experimental setup.
Software Dependencies No The paper states 'Our experiments are implemented using the pfl-research framework (Granqvist et al., 2024)', but it does not specify version numbers for this framework or any other software dependencies like Python, PyTorch, TensorFlow, or specific libraries.
Experiment Setup Yes For CIFAR10 we vary the local batch size over [10, 15, 20, 25], the local number of epochs over [1, 2, 5, 10] and the local learning rate over [0.005, 0.01, 0.05, 0.1, 0.5]. ... Global learning rate for Fed Avg is 1.0, client cohort size is 50, and the number of global training rounds is 1500. For FEMNIST we vary the local number of epochs over [1, 2, 5, 10] and the local learning rate over [0.005, 0.01, 0.05]. ... Global learning rate for Fed Avg is 1.0, client cohort size is 50, the number of global training rounds is 1500 and the local batch size is 10.