Private measurement of nonlinear correlations between data hosted across
multiple parties
- URL: http://arxiv.org/abs/2110.09670v1
- Date: Tue, 19 Oct 2021 00:31:26 GMT
- Title: Private measurement of nonlinear correlations between data hosted across
multiple parties
- Authors: Praneeth Vepakomma, Subha Nawer Pushpita, Ramesh Raskar
- Abstract summary: We introduce a differentially private method to measure nonlinear correlations between sensitive data hosted across two entities.
This work has direct applications to private feature screening, private independence testing, private k-sample tests, private multi-party causal inference and private data synthesis.
- Score: 14.93584434176082
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We introduce a differentially private method to measure nonlinear
correlations between sensitive data hosted across two entities. We provide
utility guarantees of our private estimator. Ours is the first such private
estimator of nonlinear correlations, to the best of our knowledge within a
multi-party setup. The important measure of nonlinear correlation we consider
is distance correlation. This work has direct applications to private feature
screening, private independence testing, private k-sample tests, private
multi-party causal inference and private data synthesis in addition to
exploratory data analysis. Code access: A link to publicly access the code is
provided in the supplementary file.
Related papers
- Differentially Private Covariate Balancing Causal Inference [8.133739801185271]
Differential privacy is the leading mathematical framework for privacy protection.
Our algorithm produces both point and interval estimators with statistical guarantees, such as consistency and rate optimality, under a given privacy budget.
arXiv Detail & Related papers (2024-10-18T18:02:13Z) - Empirical Mean and Frequency Estimation Under Heterogeneous Privacy: A Worst-Case Analysis [5.755004576310333]
Differential Privacy (DP) is the current gold-standard for measuring privacy.
We consider the problems of empirical mean estimation for univariate data and frequency estimation for categorical data, subject to heterogeneous privacy constraints.
We prove some optimality results, under both PAC error and mean-squared error, for our proposed algorithms and demonstrate superior performance over other baseline techniques experimentally.
arXiv Detail & Related papers (2024-07-15T22:46:02Z) - Initialization Matters: Privacy-Utility Analysis of Overparameterized
Neural Networks [72.51255282371805]
We prove a privacy bound for the KL divergence between model distributions on worst-case neighboring datasets.
We find that this KL privacy bound is largely determined by the expected squared gradient norm relative to model parameters during training.
arXiv Detail & Related papers (2023-10-31T16:13:22Z) - Differentially Private Linear Regression with Linked Data [3.9325957466009203]
Differential privacy, a mathematical notion from computer science, is a rising tool offering robust privacy guarantees.
Recent work focuses on developing differentially private versions of individual statistical and machine learning tasks.
We present two differentially private algorithms for linear regression with linked data.
arXiv Detail & Related papers (2023-08-01T21:00:19Z) - Differentially Private Confidence Intervals for Proportions under Stratified Random Sampling [14.066813980992132]
With the increase of data privacy awareness, developing a private version of confidence intervals has gained growing attention.
Recent work has been done around differentially private confidence intervals, yet rigorous methodologies on differentially private confidence intervals have not been studied.
We propose three differentially private algorithms for constructing confidence intervals for proportions under stratified random sampling.
arXiv Detail & Related papers (2023-01-19T21:25:41Z) - How Do Input Attributes Impact the Privacy Loss in Differential Privacy? [55.492422758737575]
We study the connection between the per-subject norm in DP neural networks and individual privacy loss.
We introduce a novel metric termed the Privacy Loss-Input Susceptibility (PLIS) which allows one to apportion the subject's privacy loss to their input attributes.
arXiv Detail & Related papers (2022-11-18T11:39:03Z) - Algorithms with More Granular Differential Privacy Guarantees [65.3684804101664]
We consider partial differential privacy (DP), which allows quantifying the privacy guarantee on a per-attribute basis.
In this work, we study several basic data analysis and learning tasks, and design algorithms whose per-attribute privacy parameter is smaller that the best possible privacy parameter for the entire record of a person.
arXiv Detail & Related papers (2022-09-08T22:43:50Z) - Private independence testing across two parties [21.236868468146348]
$pi$-test is a privacy-preserving algorithm for testing statistical independence between data distributed across multiple parties.
We establish both additive and multiplicative error bounds on the utility of our differentially private test.
arXiv Detail & Related papers (2022-07-08T02:13:05Z) - Mixed Differential Privacy in Computer Vision [133.68363478737058]
AdaMix is an adaptive differentially private algorithm for training deep neural network classifiers using both private and public image data.
A few-shot or even zero-shot learning baseline that ignores private data can outperform fine-tuning on a large private dataset.
arXiv Detail & Related papers (2022-03-22T06:15:43Z) - Robustness Threats of Differential Privacy [70.818129585404]
We experimentally demonstrate that networks, trained with differential privacy, in some settings might be even more vulnerable in comparison to non-private versions.
We study how the main ingredients of differentially private neural networks training, such as gradient clipping and noise addition, affect the robustness of the model.
arXiv Detail & Related papers (2020-12-14T18:59:24Z) - Towards practical differentially private causal graph discovery [74.7791110594082]
Causal graph discovery refers to the process of discovering causal relation graphs from purely observational data.
We present a differentially private causal graph discovery algorithm, Priv-PC, which improves both utility and running time compared to the state-of-the-art.
arXiv Detail & Related papers (2020-06-15T18:30:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.