Notice: The reproducibility variables underlying each score are classified using an automated LLM-based pipeline, validated against a manually labeled dataset. LLM-based classification introduces uncertainty and potential bias; scores should be interpreted as estimates. Full accuracy metrics and methodology are described in [1].
Privacy Attacks in Decentralized Learning
Authors: Abdellah El Mrini, Edwige Cyffers, Aurรฉlien Bellet
ICML 2024 | Venue PDF | LLM Run Details
| Reproducibility Variable | Result | LLM Response |
|---|---|---|
| Research Type | Experimental | We empirically evaluate the performance of our attacks on synthetic and real graphs and datasets. Our attacks prove to be effective in practice across various graph structures. |
| Researcher Affiliation | Academia | 1School of Computer and Communication Sciences, EPFL, Switzerland 2Universit e de Lille, Inria, CNRS, Centrale Lille, UMR 9189 CRISt AL, F-59000 Lille, France 3Inria, Univ Montpellier, Montpellier, France. |
| Pseudocode | Yes | Algorithm 1 Knowledge matrix construction; Algorithm 2 Building the knowledge matrix for D-GD; Algorithm 3 Removing the attackers contributions; Algorithm 4 Building the covariance matrix |
| Open Source Code | Yes | Our code is available at https://github.com/Abdellah Elmrini/dec Attack. |
| Open Datasets | Yes | We first focus on the Cifar10 dataset (Krizhevsky, 2009); We use a small convolutional neural network (see details in Appendix G) on the MNIST dataset |
| Dataset Splits | No | The paper mentions using CIFAR10 and MNIST datasets but does not provide specific details about the training, validation, or test splits (e.g., percentages, sample counts, or explicit references to standard splits). |
| Hardware Specification | No | Experiments presented in this paper were carried out using the Grid 5000 testbed, supported by a scientific interest group hosted by Inria and including CNRS, RENATER and several Universities as well as other organizations (see https://www.grid5000.fr). |
| Software Dependencies | No | The paper mentions using a "fully connected layer with a softmax activation" and a "small convolutional neural network" but does not specify software dependencies with version numbers (e.g., Python, PyTorch, TensorFlow versions). |
| Experiment Setup | Yes | We start running our attack when the model is close to convergence so that gradients are more stable. [...] The learning rate plays an important role: it should be small enough to ensure that gradients do not vary wildly across iterations. We illustrate this behavior in Appendix H. [...] logistic regression, learning rate 10^-4; [...] convnet, learning rate 10^-6 |