Private Learning of Halfspaces: Simplifying the Construction and Reducing the Sample Complexity

Authors: Haim Kaplan, Yishay Mansour, Uri Stemmer, Eliad Tsfadia

NeurIPS 2020 | Conference PDF | Archive PDF | Plain Text | LLM Run Details

Reproducibility Variable Result LLM Response
Research Type Theoretical We present a differentially private learner for halfspaces over a finite grid G in Rd with sample complexity d2.5 2log |G|, which improves the state-of-the-art result of [Beimel et al., COLT 2019] by a d2 factor. The building block for our learner is a new differentially private algorithm for approximately solving the linear feasibility problem: Given a feasible collection of m linear constraints of the form Ax b, the task is to privately identify a solution x that satisfies most of the constraints. Our algorithm is iterative, where each iteration determines the next coordinate of the constructed solution x. The paper focuses on theoretical contributions, algorithm design, and proofs related to sample complexity, which is characteristic of theoretical research.
Researcher Affiliation Collaboration Haim Kaplan Tel Aviv University and Google Research haimk@tau.ac.il Yishay Mansour Tel Aviv University and Google Research mansour.yishay@gmail.com Uri Stemmer Ben-Gurion University and Google Research u@uri.co.il Eliad Tsfadia Tel Aviv University and Google Research eliadtsfadia@gmail.com
Pseudocode Yes Figure 1: Algorithm AOptimize High Dim Func (i) Let α, β, ε, δ (0, 1) be the utility/privacy parameters, let S X be an input dataset, let n Xi( ) od i=1 be an iterative sequence of finite domains, and let 1. (ii) For i = 1 to d do: (a) Let Qx 1,...,x i 1 be the function from Definition 3.3. (b) Let Xi = Xi(x 1, . . . , x i 1). (c) Execute ARec Concave with the function Qx 1,...,x i 1, domain Xi, and parameters: r = (1 α 2d )i 1MS, α = α 2d , β = β d , ε = ε 2 2d ln(2/δ), δ = δ 2d. Let x i be its output. (iii) Return x = (x 1, . . . , x d).
Open Source Code No The paper does not provide any concrete statement about releasing its own source code, nor does it provide a link to a repository for the described methodology.
Open Datasets No This is a theoretical paper focusing on algorithm design and sample complexity. It does not use specific real-world datasets for training or provide information on their public availability.
Dataset Splits No This is a theoretical paper. It does not describe any empirical experiments, and therefore no training, validation, or test dataset splits are provided.
Hardware Specification No This is a theoretical paper. It does not discuss any specific hardware used for computations or experiments.
Software Dependencies No This is a theoretical paper. It does not mention any specific software components or their version numbers that would be required to replicate its findings.
Experiment Setup No This is a theoretical paper. It does not describe any empirical experimental setup details or hyperparameters.