Weakly-Private Information Retrieval From MDS-Coded Distributed Storage
- URL: http://arxiv.org/abs/2401.09412v1
- Date: Wed, 17 Jan 2024 18:51:04 GMT
- Title: Weakly-Private Information Retrieval From MDS-Coded Distributed Storage
- Authors: Asbjørn O. Orvedal, Hsuan-Yin Lin, Eirik Rosnes,
- Abstract summary: In WPIR, a user wishes to retrieve a piece of data from a set of servers without leaking too much information about which piece of data she is interested in.
We study and provide the first WPIR protocols for this scenario and present results on their optimal trade-off between download rate and information leakage using the maximal leakage privacy metric.
- Score: 11.955988388140725
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: We consider the problem of weakly-private information retrieval (WPIR) when data is encoded by a maximum distance separable code and stored across multiple servers. In WPIR, a user wishes to retrieve a piece of data from a set of servers without leaking too much information about which piece of data she is interested in. We study and provide the first WPIR protocols for this scenario and present results on their optimal trade-off between download rate and information leakage using the maximal leakage privacy metric.
Related papers
- LLM-PBE: Assessing Data Privacy in Large Language Models [111.58198436835036]
Large Language Models (LLMs) have become integral to numerous domains, significantly advancing applications in data management, mining, and analysis.
Despite the critical nature of this issue, there has been no existing literature to offer a comprehensive assessment of data privacy risks in LLMs.
Our paper introduces LLM-PBE, a toolkit crafted specifically for the systematic evaluation of data privacy risks in LLMs.
arXiv Detail & Related papers (2024-08-23T01:37:29Z) - Robust Utility-Preserving Text Anonymization Based on Large Language Models [80.5266278002083]
Text anonymization is crucial for sharing sensitive data while maintaining privacy.
Existing techniques face the emerging challenges of re-identification attack ability of Large Language Models.
This paper proposes a framework composed of three LLM-based components -- a privacy evaluator, a utility evaluator, and an optimization component.
arXiv Detail & Related papers (2024-07-16T14:28:56Z) - Secure Distributed Storage: Optimal Trade-Off Between Storage Rate and Privacy Leakage [1.6881346757176976]
We consider the problem of storing data in a distributed manner over $T$ servers.
Specifically, the data needs to (i) be recoverable from any $tau$ servers, and (ii) remain private from any $z$ colluding servers.
For this model, our main results are (i) the fundamental trade-off between storage size and the level of desired privacy, and (ii) the optimal amount of local randomness necessary at the encoder.
arXiv Detail & Related papers (2024-03-15T20:50:46Z) - Quantum Private Information Retrieval from Coded Storage Systems [0.0]
In a Quantum Private Information Retrieval (QPIR) protocol, a user can retrieve information from a database by downloading quantum systems from multiple servers.
This thesis aims to develop QPIR protocols for coded storage by combining known classical PIR protocols with quantum communication algorithms.
arXiv Detail & Related papers (2023-12-10T12:36:10Z) - Outsourcing Training without Uploading Data via Efficient Collaborative
Open-Source Sampling [49.87637449243698]
Traditional outsourcing requires uploading device data to the cloud server.
We propose to leverage widely available open-source data, which is a massive dataset collected from public and heterogeneous sources.
We develop a novel strategy called Efficient Collaborative Open-source Sampling (ECOS) to construct a proximal proxy dataset from open-source data for cloud training.
arXiv Detail & Related papers (2022-10-23T00:12:18Z) - Scalable Neural Data Server: A Data Recommender for Transfer Learning [70.06289658553675]
Transfer learning is a popular strategy for leveraging additional data to improve the downstream performance.
Nerve Data Server (NDS), a search engine that recommends relevant data for a given downstream task, has been previously proposed to address this problem.
NDS uses a mixture of experts trained on data sources to estimate similarity between each source and the downstream task.
SNDS represents both data sources and downstream tasks by their proximity to the intermediary datasets.
arXiv Detail & Related papers (2022-06-19T12:07:32Z) - Robbing the Fed: Directly Obtaining Private Data in Federated Learning
with Modified Models [56.0250919557652]
Federated learning has quickly gained popularity with its promises of increased user privacy and efficiency.
Previous attacks on user privacy have been limited in scope and do not scale to gradient updates aggregated over even a handful of data points.
We introduce a new threat model based on minimal but malicious modifications of the shared model architecture.
arXiv Detail & Related papers (2021-10-25T15:52:06Z) - Generative Adversarial User Privacy in Lossy Single-Server Information
Retrieval [18.274573259364026]
We consider the problem of information retrieval from a dataset of files stored on a single server under both a user distortion and a user privacy constraint.
Specifically, a user requesting a file from the dataset should be able to reconstruct the requested file with a prescribed distortion.
In addition, the identity of the requested file should be kept private from the server with a prescribed privacy level.
arXiv Detail & Related papers (2020-12-07T18:31:51Z) - Deep Directed Information-Based Learning for Privacy-Preserving Smart
Meter Data Release [30.409342804445306]
We study the problem in the context of time series data and smart meters (SMs) power consumption measurements.
We introduce the Directed Information (DI) as a more meaningful measure of privacy in the considered setting.
Our empirical studies on real-world data sets from SMs measurements in the worst-case scenario show the existing trade-offs between privacy and utility.
arXiv Detail & Related papers (2020-11-20T13:41:11Z) - GDPR: When the Right to Access Personal Data Becomes a Threat [63.732639864601914]
We examine more than 300 data controllers performing for each of them a request to access personal data.
We find that 50.4% of the data controllers that handled the request, have flaws in the procedure of identifying the users.
With the undesired and surprising result that, in its present deployment, has actually decreased the privacy of the users of web services.
arXiv Detail & Related papers (2020-05-04T22:01:46Z) - Quantum Private Information Retrieval from Coded and Colluding Servers [16.23970875497387]
In the quantum PIR (QPIR) setting, a user privately retrieves a classical file by receiving quantum information from the servers.
In this paper, the QPIR setting is extended to account for maximum distance separable (MDS) coded servers.
The rates achieved are better than those known or conjectured in the classical counterparts.
arXiv Detail & Related papers (2020-01-16T15:19:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.