Generative Adversarial User Privacy in Lossy Single-Server Information
Retrieval
- URL: http://arxiv.org/abs/2012.03902v1
- Date: Mon, 7 Dec 2020 18:31:51 GMT
- Title: Generative Adversarial User Privacy in Lossy Single-Server Information
Retrieval
- Authors: Chung-Wei Weng, Yauhen Yakimenka, Hsuan-Yin Lin, Eirik Rosnes, Joerg
Kliewer
- Abstract summary: We consider the problem of information retrieval from a dataset of files stored on a single server under both a user distortion and a user privacy constraint.
Specifically, a user requesting a file from the dataset should be able to reconstruct the requested file with a prescribed distortion.
In addition, the identity of the requested file should be kept private from the server with a prescribed privacy level.
- Score: 18.274573259364026
- License: http://creativecommons.org/publicdomain/zero/1.0/
- Abstract: We consider the problem of information retrieval from a dataset of files
stored on a single server under both a user distortion and a user privacy
constraint. Specifically, a user requesting a file from the dataset should be
able to reconstruct the requested file with a prescribed distortion, and in
addition, the identity of the requested file should be kept private from the
server with a prescribed privacy level. The proposed model can be seen as an
extension of the well-known concept of private information retrieval by
allowing for distortion in the retrieval process and relaxing the perfect
privacy requirement. We initiate the study of the tradeoff between download
rate, distortion, and user privacy leakage, and show that the optimal
rate-distortion-leakage tradeoff is convex and that in the limit of large file
sizes this allows for a concise information-theoretical formulation in terms of
mutual information. Moreover, we propose a new data-driven framework by
leveraging recent advancements in generative adversarial models which allows a
user to learn efficient schemes in terms of download rate from the data itself.
Learning the scheme is formulated as a constrained minimax game between a user
which desires to keep the identity of the requested file private and an
adversary that tries to infer which file the user is interested in under a
distortion constraint. In general, guaranteeing a certain privacy level leads
to a higher rate-distortion tradeoff curve, and hence a sacrifice in either
download rate or distortion. We evaluate the performance of the scheme on a
synthetic Gaussian dataset as well as on the MNIST and CIFAR-$10$ datasets. For
the MNIST dataset, the data-driven approach significantly outperforms a
proposed general achievable scheme combining source coding with the download of
multiple files, while for CIFAR-$10$ the performances are comparable.
Related papers
- Robust Utility-Preserving Text Anonymization Based on Large Language Models [80.5266278002083]
Text anonymization is crucial for sharing sensitive data while maintaining privacy.
Existing techniques face the emerging challenges of re-identification attack ability of Large Language Models.
This paper proposes a framework composed of three LLM-based components -- a privacy evaluator, a utility evaluator, and an optimization component.
arXiv Detail & Related papers (2024-07-16T14:28:56Z) - Federated Face Forgery Detection Learning with Personalized Representation [63.90408023506508]
Deep generator technology can produce high-quality fake videos that are indistinguishable, posing a serious social threat.
Traditional forgery detection methods directly centralized training on data.
The paper proposes a novel federated face forgery detection learning with personalized representation.
arXiv Detail & Related papers (2024-06-17T02:20:30Z) - $\alpha$-Mutual Information: A Tunable Privacy Measure for Privacy
Protection in Data Sharing [4.475091558538915]
This paper adopts Arimoto's $alpha$-Mutual Information as a tunable privacy measure.
We formulate a general distortion-based mechanism that manipulates the original data to offer privacy protection.
arXiv Detail & Related papers (2023-10-27T16:26:14Z) - Private Set Generation with Discriminative Information [63.851085173614]
Differentially private data generation is a promising solution to the data privacy challenge.
Existing private generative models are struggling with the utility of synthetic samples.
We introduce a simple yet effective method that greatly improves the sample utility of state-of-the-art approaches.
arXiv Detail & Related papers (2022-11-07T10:02:55Z) - Smooth Anonymity for Sparse Graphs [69.1048938123063]
differential privacy has emerged as the gold standard of privacy, however, when it comes to sharing sparse datasets.
In this work, we consider a variation of $k$-anonymity, which we call smooth-$k$-anonymity, and design simple large-scale algorithms that efficiently provide smooth-$k$-anonymity.
arXiv Detail & Related papers (2022-07-13T17:09:25Z) - Group privacy for personalized federated learning [4.30484058393522]
Federated learning is a type of collaborative machine learning, where participating clients process their data locally, sharing only updates to the collaborative model.
We propose a method to provide group privacy guarantees exploiting some key properties of $d$-privacy.
arXiv Detail & Related papers (2022-06-07T15:43:45Z) - Mixed Differential Privacy in Computer Vision [133.68363478737058]
AdaMix is an adaptive differentially private algorithm for training deep neural network classifiers using both private and public image data.
A few-shot or even zero-shot learning baseline that ignores private data can outperform fine-tuning on a large private dataset.
arXiv Detail & Related papers (2022-03-22T06:15:43Z) - Privacy-Aware Compression for Federated Data Analysis [31.970815289473965]
Federated data analytics is a framework for distributed data analysis where a server compiles noisy responses from a group of low-bandwidth user devices to estimate aggregate statistics.
Two major challenges in this framework are privacy, since user data is often sensitive, and compression, since the user devices have low network bandwidth.
We take a holistic look at the problem and design a family of privacy-aware compression mechanisms that work for any given communication budget.
arXiv Detail & Related papers (2022-03-15T17:57:13Z) - Robbing the Fed: Directly Obtaining Private Data in Federated Learning
with Modified Models [56.0250919557652]
Federated learning has quickly gained popularity with its promises of increased user privacy and efficiency.
Previous attacks on user privacy have been limited in scope and do not scale to gradient updates aggregated over even a handful of data points.
We introduce a new threat model based on minimal but malicious modifications of the shared model architecture.
arXiv Detail & Related papers (2021-10-25T15:52:06Z) - Federated Deep Learning with Bayesian Privacy [28.99404058773532]
Federated learning (FL) aims to protect data privacy by cooperatively learning a model without sharing private data among users.
Homomorphic encryption (HE) based methods provide secure privacy protections but suffer from extremely high computational and communication overheads.
Deep learning with Differential Privacy (DP) was implemented as a practical learning algorithm at a manageable cost in complexity.
arXiv Detail & Related papers (2021-09-27T12:48:40Z) - Generating private data with user customization [9.415164800448853]
Mobile devices can produce and store large amounts of data that can enhance machine learning models.
However, this data may contain private information specific to the data owner that prevents the release of the data.
We want to reduce the correlation between user-specific private information and the data while retaining the useful information.
arXiv Detail & Related papers (2020-12-02T19:13:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.