Privacy-preserving Data Sharing on Vertically Partitioned Data
- URL: http://arxiv.org/abs/2010.09293v2
- Date: Fri, 2 Sep 2022 07:59:04 GMT
- Title: Privacy-preserving Data Sharing on Vertically Partitioned Data
- Authors: Razane Tajeddine, Joonas J\"alk\"o, Samuel Kaski, and Antti Honkela
- Abstract summary: We introduce a differentially private method for generating synthetic data from vertically partitioned data.
We train a mixture model over partitioned data using variational inference.
We rigorously define the privacy guarantees with respect to the different players in the system.
- Score: 16.167363414383576
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this work, we introduce a differentially private method for generating
synthetic data from vertically partitioned data, \emph{i.e.}, where data of the
same individuals is distributed across multiple data holders or parties. We
present a differentially privacy stochastic gradient descent (DP-SGD) algorithm
to train a mixture model over such partitioned data using variational
inference. We modify a secure multiparty computation (MPC) framework to combine
MPC with differential privacy (DP), in order to use differentially private MPC
effectively to learn a probabilistic generative model under DP on such
vertically partitioned data.
Assuming the mixture components contain no dependencies across different
parties, the objective function can be factorized into a sum of products of the
contributions calculated by the parties. Finally, MPC is used to compute the
aggregate between the different contributions. Moreover, we rigorously define
the privacy guarantees with respect to the different players in the system. To
demonstrate the accuracy of our method, we run our algorithm on the Adult
dataset from the UCI machine learning repository, where we obtain comparable
results to the non-partitioned case.
Related papers
- Differentially Private Random Feature Model [52.468511541184895]
We produce a differentially private random feature model for privacy-preserving kernel machines.
We show that our method preserves privacy and derive a generalization error bound for the method.
arXiv Detail & Related papers (2024-12-06T05:31:08Z) - Differential Privacy on Trust Graphs [54.55190841518906]
We study differential privacy (DP) in a multi-party setting where each party only trusts a (known) subset of the other parties with its data.
We give a DP algorithm for aggregation with a much better privacy-utility trade-off than in the well-studied local model of DP.
arXiv Detail & Related papers (2024-10-15T20:31:04Z) - Enhancing Scalability of Metric Differential Privacy via Secret Dataset Partitioning and Benders Decomposition [1.283608820493284]
Metric Differential Privacy (mDP) extends the concept of Differential Privacy (DP) to serve as a new paradigm of data.
It is designed to protect secret data represented in general metric space, such as text data encoded as word embeddings or geo-location data on the road network or grid maps.
arXiv Detail & Related papers (2024-05-07T14:19:09Z) - Collaborative Heterogeneous Causal Inference Beyond Meta-analysis [68.4474531911361]
We propose a collaborative inverse propensity score estimator for causal inference with heterogeneous data.
Our method shows significant improvements over the methods based on meta-analysis when heterogeneity increases.
arXiv Detail & Related papers (2024-04-24T09:04:36Z) - Noise Variance Optimization in Differential Privacy: A Game-Theoretic Approach Through Per-Instance Differential Privacy [7.264378254137811]
Differential privacy (DP) can measure privacy loss by observing the changes in the distribution caused by the inclusion of individuals in the target dataset.
DP has been prominent in safeguarding datasets in machine learning in industry giants like Apple and Google.
We propose per-instance DP (pDP) as a constraint, measuring privacy loss for each data instance and optimizing noise tailored to individual instances.
arXiv Detail & Related papers (2024-04-24T06:51:16Z) - Incentives in Private Collaborative Machine Learning [56.84263918489519]
Collaborative machine learning involves training models on data from multiple parties.
We introduce differential privacy (DP) as an incentive.
We empirically demonstrate the effectiveness and practicality of our approach on synthetic and real-world datasets.
arXiv Detail & Related papers (2024-04-02T06:28:22Z) - Personalized Graph Federated Learning with Differential Privacy [6.282767337715445]
This paper presents a personalized graph federated learning (PGFL) framework in which distributedly connected servers and their respective edge devices collaboratively learn device or cluster-specific models.
We study a variant of the PGFL implementation that utilizes differential privacy, specifically zero-concentrated differential privacy, where a noise sequence perturbs model exchanges.
Our analysis shows that the algorithm ensures local differential privacy for all clients in terms of zero-concentrated differential privacy.
arXiv Detail & Related papers (2023-06-10T09:52:01Z) - Privacy-Preserving and Lossless Distributed Estimation of
High-Dimensional Generalized Additive Mixed Models [0.9023847175654603]
We propose an algorithm for a distributed, privacy-preserving, and lossless estimation of generalized additive mixed models (GAMM) using component-wise gradient boosting (CWB)
Our adaption of CWB preserves all the important properties of the original algorithm, such as an unbiased feature selection and the feasibility to fit models in high-dimensional feature spaces.
We also showcase the efficacy of our algorithm on a distributed heart disease data set and compare it with state-of-the-art methods.
arXiv Detail & Related papers (2022-10-14T11:41:18Z) - Differentially Private Federated Learning with Laplacian Smoothing [72.85272874099644]
Federated learning aims to protect data privacy by collaboratively learning a model without sharing private data among users.
An adversary may still be able to infer the private training data by attacking the released model.
Differential privacy provides a statistical protection against such attacks at the price of significantly degrading the accuracy or utility of the trained models.
arXiv Detail & Related papers (2020-05-01T04:28:38Z) - Secure Metric Learning via Differential Pairwise Privacy [36.946123614592054]
This paper studies, for the first time, how pairwise information can be leaked to attackers during distance metric learning.
We develop differential pairwise privacy (DPP), generalizing the definition of standard differential privacy, for secure metric learning.
arXiv Detail & Related papers (2020-03-30T12:47:48Z) - User-Level Privacy-Preserving Federated Learning: Analysis and
Performance Optimization [77.43075255745389]
Federated learning (FL) is capable of preserving private data from mobile terminals (MTs) while training the data into useful models.
From a viewpoint of information theory, it is still possible for a curious server to infer private information from the shared models uploaded by MTs.
We propose a user-level differential privacy (UDP) algorithm by adding artificial noise to the shared models before uploading them to servers.
arXiv Detail & Related papers (2020-02-29T10:13:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.