Lossless Compression of Efficient Private Local Randomizers
- URL: http://arxiv.org/abs/2102.12099v1
- Date: Wed, 24 Feb 2021 07:04:30 GMT
- Title: Lossless Compression of Efficient Private Local Randomizers
- Authors: Vitaly Feldman and Kunal Talwar
- Abstract summary: Locally Differentially Private (LDP) Reports are commonly used for collection of statistics and machine learning in the federated setting.
In many cases the best known LDP algorithms require sending prohibitively large messages from the client device to the server.
This has led to significant efforts on reducing the communication cost of LDP algorithms.
- Score: 55.657133416044104
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Locally Differentially Private (LDP) Reports are commonly used for collection
of statistics and machine learning in the federated setting. In many cases the
best known LDP algorithms require sending prohibitively large messages from the
client device to the server (such as when constructing histograms over large
domain or learning a high-dimensional model). This has led to significant
efforts on reducing the communication cost of LDP algorithms.
At the same time LDP reports are known to have relatively little information
about the user's data due to randomization. Several schemes are known that
exploit this fact to design low-communication versions of LDP algorithm but all
of them do so at the expense of a significant loss in utility. Here we
demonstrate a general approach that, under standard cryptographic assumptions,
compresses every efficient LDP algorithm with negligible loss in privacy and
utility guarantees. The practical implication of our result is that in typical
applications the message can be compressed to the size of the server's
pseudo-random generator seed. More generally, we relate the properties of an
LDP randomizer to the power of a pseudo-random generator that suffices for
compressing the LDP randomizer. From this general approach we derive
low-communication algorithms for the problems of frequency estimation and
high-dimensional mean estimation. Our algorithms are simpler and more accurate
than existing low-communication LDP algorithms for these well-studied problems.
Related papers
- Sketches-based join size estimation under local differential privacy [3.0945730947183203]
Join size estimation on sensitive data poses a risk of privacy leakage.
Local differential privacy (LDP) is a solution to preserve privacy while collecting sensitive data.
We introduce a novel algorithm called LDPJoinSketch for sketch-based join size estimation under LDP.
arXiv Detail & Related papers (2024-05-19T01:21:54Z) - Noise Variance Optimization in Differential Privacy: A Game-Theoretic Approach Through Per-Instance Differential Privacy [7.264378254137811]
Differential privacy (DP) can measure privacy loss by observing the changes in the distribution caused by the inclusion of individuals in the target dataset.
DP has been prominent in safeguarding datasets in machine learning in industry giants like Apple and Google.
We propose per-instance DP (pDP) as a constraint, measuring privacy loss for each data instance and optimizing noise tailored to individual instances.
arXiv Detail & Related papers (2024-04-24T06:51:16Z) - Semidefinite programs simulate approximate message passing robustly [3.1528188401664137]
Approximate message passing (AMP) is a family of iterative algorithms that generalize power.
AMP algorithms are known to optimally solve many average-case optimization problems.
arXiv Detail & Related papers (2023-11-15T15:00:48Z) - Fast Optimal Locally Private Mean Estimation via Random Projections [58.603579803010796]
We study the problem of locally private mean estimation of high-dimensional vectors in the Euclidean ball.
We propose a new algorithmic framework, ProjUnit, for private mean estimation.
Our framework is deceptively simple: each randomizer projects its input to a random low-dimensional subspace, normalizes the result, and then runs an optimal algorithm.
arXiv Detail & Related papers (2023-06-07T14:07:35Z) - Private and Communication-Efficient Algorithms for Entropy Estimation [21.195965074919602]
We give improved private and communication-efficient algorithms for estimating several popular measures of the entropy of a distribution.
All of our algorithms have constant communication cost and satisfy local differential privacy.
arXiv Detail & Related papers (2023-05-12T20:35:10Z) - Communication-Efficient Adam-Type Algorithms for Distributed Data Mining [93.50424502011626]
We propose a class of novel distributed Adam-type algorithms (emphi.e., SketchedAMSGrad) utilizing sketching.
Our new algorithm achieves a fast convergence rate of $O(frac1sqrtnT + frac1(k/d)2 T)$ with the communication cost of $O(k log(d))$ at each iteration.
arXiv Detail & Related papers (2022-10-14T01:42:05Z) - RL for Latent MDPs: Regret Guarantees and a Lower Bound [74.41782017817808]
We consider the regret problem for reinforcement learning in latent Markov Decision Processes (LMDP)
In an LMDP, an MDP is randomly drawn from a set of $M$ possible MDPs at the beginning of the interaction, but the identity of the chosen MDP is not revealed to the agent.
We show that the key link is a notion of separation between the MDP system dynamics.
arXiv Detail & Related papers (2021-02-09T16:49:58Z) - A One-Pass Private Sketch for Most Machine Learning Tasks [48.17461258268463]
Differential privacy (DP) is a compelling privacy definition that explains the privacy-utility tradeoff via formal, provable guarantees.
We propose a private sketch that supports a multitude of machine learning tasks including regression, classification, density estimation, and more.
Our sketch consists of randomized contingency tables that are indexed with locality-sensitive hashing and constructed with an efficient one-pass algorithm.
arXiv Detail & Related papers (2020-06-16T17:47:48Z) - User-Level Privacy-Preserving Federated Learning: Analysis and
Performance Optimization [77.43075255745389]
Federated learning (FL) is capable of preserving private data from mobile terminals (MTs) while training the data into useful models.
From a viewpoint of information theory, it is still possible for a curious server to infer private information from the shared models uploaded by MTs.
We propose a user-level differential privacy (UDP) algorithm by adding artificial noise to the shared models before uploading them to servers.
arXiv Detail & Related papers (2020-02-29T10:13:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.