Notice: The reproducibility variables underlying each score are classified using an automated LLM-based pipeline, validated against a manually labeled dataset. LLM-based classification introduces uncertainty and potential bias; scores should be interpreted as estimates. Full accuracy metrics and methodology are described in [1].
Minimax Optimal Estimation of Approximate Differential Privacy on Neighboring Databases
Authors: Xiyang Liu, Sewoong Oh
NeurIPS 2019 | Venue PDF | LLM Run Details
| Reproducibility Variable | Result | LLM Response |
|---|---|---|
| Research Type | Experimental | We present numerical experiments supporting our theoretical predictions in Section 3. Figure 1 (a) illustrates the Mean Square Error (MSE) for estimating dε(P%Q) between uniform distribution P and Zipf distribution Q... We demonstrate how we can use Algorithm 2 to detect mechanisms with false claim of DP guarantees on four types of mechanisms |
| Researcher Affiliation | Academia | Xiyang Liu Sewoong Oh Allen School of Computer Science and Engineering, University of Washington EMAIL |
| Pseudocode | Yes | Algorithm 1 Differential Privacy (DP) estimator with known P ... Algorithm 2 Differential Privacy (DP) estimator |
| Open Source Code | Yes | We present the experiment details in Appendix B and the code to reproduce our experiments at https://github.com/xiyangl3/adp-estimator. |
| Open Datasets | No | The paper mentions using 'uniform distribution P and Zipf distribution Q' (synthetic data) and 'real-world experiments' but does not provide any concrete access information (link, DOI, citation with author/year) for these or any other publicly available datasets. |
| Dataset Splits | No | The paper does not explicitly provide details about training/validation/test dataset splits, such as percentages, absolute counts, or references to predefined splits for model training. |
| Hardware Specification | Yes | All experiments were performed on a machine with Intel Core i7-4770K CPU (3.50GHz) and 32GB RAM. |
| Software Dependencies | No | The paper mentions 'Python 3.6' with a version but does not provide version numbers for other key software components like 'numpy', thus not fulfilling the requirement of multiple versioned components or all mentioned components being versioned. |
| Experiment Setup | No | The paper does not provide specific experimental setup details such as hyperparameters (e.g., learning rate, batch size, number of epochs) or optimizer settings. It refers to Appendix B.2 for settings, but this appendix discusses the mechanisms rather than detailed training configurations. |