Discrete Distribution Estimation under User-level Local Differential
Privacy
- URL: http://arxiv.org/abs/2211.03757v1
- Date: Mon, 7 Nov 2022 18:29:32 GMT
- Title: Discrete Distribution Estimation under User-level Local Differential
Privacy
- Authors: Jayadev Acharya, Yuhan Liu, Ziteng Sun
- Abstract summary: We study discrete distribution estimation under user-level local differential privacy (LDP)
In user-level $varepsilon$-LDP, each user has $mge1$ samples and the privacy of all $m$ samples must be preserved simultaneously.
- Score: 37.65849910114053
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We study discrete distribution estimation under user-level local differential
privacy (LDP). In user-level $\varepsilon$-LDP, each user has $m\ge1$ samples
and the privacy of all $m$ samples must be preserved simultaneously. We resolve
the following dilemma: While on the one hand having more samples per user
should provide more information about the underlying distribution, on the other
hand, guaranteeing the privacy of all $m$ samples should make the estimation
task more difficult. We obtain tight bounds for this problem under almost all
parameter regimes. Perhaps surprisingly, we show that in suitable parameter
regimes, having $m$ samples per user is equivalent to having $m$ times more
users, each with only one sample. Our results demonstrate interesting phase
transitions for $m$ and the privacy parameter $\varepsilon$ in the estimation
risk. Finally, connecting with recent results on shuffled DP, we show that
combined with random shuffling, our algorithm leads to optimal error guarantees
(up to logarithmic factors) under the central model of user-level DP in certain
parameter regimes. We provide several simulations to verify our theoretical
findings.
Related papers
- Distribution-Aware Mean Estimation under User-level Local Differential Privacy [5.267844649650687]
We consider the problem of mean estimation under user-level local differential privacy, where $n$ users are contributing through their local pool of data samples.
Based on a distribution-aware mean estimation algorithm, we establish an $M$-dependent upper bounds on the worst-case risk over $mu$ for the task of mean estimation.
arXiv Detail & Related papers (2024-10-12T11:57:52Z) - Private Mean Estimation with Person-Level Differential Privacy [6.621676316292624]
We study person-level differentially private mean estimation in the case where each person holds multiple samples.
We give computationally efficient algorithms under approximate-DP and computationally inefficient algorithms under pure DP, and our nearly matching lower bounds hold for the most permissive case of approximate DP.
arXiv Detail & Related papers (2024-05-30T18:20:35Z) - Tight and Robust Private Mean Estimation with Few Users [16.22135057266913]
We study high-dimensional mean estimation under user-level differential privacy.
We design an $(eps,delta)$-differentially private mechanism using as few users as possible.
arXiv Detail & Related papers (2021-10-22T16:02:21Z) - User-Level Private Learning via Correlated Sampling [49.453751858361265]
We consider the setting where each user holds $m$ samples and the privacy protection is enforced at the level of each user's data.
We show that, in this setting, we may learn with a much fewer number of users.
arXiv Detail & Related papers (2021-10-21T15:33:53Z) - Learning with User-Level Privacy [61.62978104304273]
We analyze algorithms to solve a range of learning tasks under user-level differential privacy constraints.
Rather than guaranteeing only the privacy of individual samples, user-level DP protects a user's entire contribution.
We derive an algorithm that privately answers a sequence of $K$ adaptively chosen queries with privacy cost proportional to $tau$, and apply it to solve the learning tasks we consider.
arXiv Detail & Related papers (2021-02-23T18:25:13Z) - Hiding Among the Clones: A Simple and Nearly Optimal Analysis of Privacy
Amplification by Shuffling [49.43288037509783]
We show that random shuffling amplifies differential privacy guarantees of locally randomized data.
Our result is based on a new approach that is simpler than previous work and extends to approximate differential privacy with nearly the same guarantees.
arXiv Detail & Related papers (2020-12-23T17:07:26Z) - Private Stochastic Non-Convex Optimization: Adaptive Algorithms and
Tighter Generalization Bounds [72.63031036770425]
We propose differentially private (DP) algorithms for bound non-dimensional optimization.
We demonstrate two popular deep learning methods on the empirical advantages over standard gradient methods.
arXiv Detail & Related papers (2020-06-24T06:01:24Z) - Locally Private Hypothesis Selection [96.06118559817057]
We output a distribution from $mathcalQ$ whose total variation distance to $p$ is comparable to the best such distribution.
We show that the constraint of local differential privacy incurs an exponential increase in cost.
Our algorithms result in exponential improvements on the round complexity of previous methods.
arXiv Detail & Related papers (2020-02-21T18:30:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.