Differentially Private Estimation of Heterogeneous Causal Effects
- URL: http://arxiv.org/abs/2202.11043v1
- Date: Tue, 22 Feb 2022 17:21:18 GMT
- Title: Differentially Private Estimation of Heterogeneous Causal Effects
- Authors: Fengshi Niu, Harsha Nori, Brian Quistorff, Rich Caruana, Donald Ngwe,
Aadharsh Kannan
- Abstract summary: We introduce a general meta-algorithm for estimating conditional average treatment effects (CATE) with differential privacy guarantees.
Our meta-algorithm can work with simple, single-stage CATE estimators such as S-learner and more complex multi-stage estimators such as DR and R-learner.
- Score: 9.355532300027727
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Estimating heterogeneous treatment effects in domains such as healthcare or
social science often involves sensitive data where protecting privacy is
important. We introduce a general meta-algorithm for estimating conditional
average treatment effects (CATE) with differential privacy (DP) guarantees. Our
meta-algorithm can work with simple, single-stage CATE estimators such as
S-learner and more complex multi-stage estimators such as DR and R-learner. We
perform a tight privacy analysis by taking advantage of sample splitting in our
meta-algorithm and the parallel composition property of differential privacy.
In this paper, we implement our approach using DP-EBMs as the base learner.
DP-EBMs are interpretable, high-accuracy models with privacy guarantees, which
allow us to directly observe the impact of DP noise on the learned causal
model. Our experiments show that multi-stage CATE estimators incur larger
accuracy loss than single-stage CATE or ATE estimators and that most of the
accuracy loss from differential privacy is due to an increase in variance, not
biased estimates of treatment effects.
Related papers
- Stratified Prediction-Powered Inference for Hybrid Language Model Evaluation [62.2436697657307]
Prediction-powered inference (PPI) is a method that improves statistical estimates based on limited human-labeled data.
We propose a method called Stratified Prediction-Powered Inference (StratPPI)
We show that the basic PPI estimates can be considerably improved by employing simple data stratification strategies.
arXiv Detail & Related papers (2024-06-06T17:37:39Z) - A Simple and Practical Method for Reducing the Disparate Impact of
Differential Privacy [21.098175634158043]
Differentially private (DP) mechanisms have been deployed in a variety of high-impact social settings.
The impact of DP on utility can vary significantly among different sub-populations.
A simple way to reduce this disparity is with stratification.
arXiv Detail & Related papers (2023-12-18T21:19:35Z) - Adaptive Differentially Quantized Subspace Perturbation (ADQSP): A Unified Framework for Privacy-Preserving Distributed Average Consensus [6.364764301218972]
We propose a general approach named adaptive differentially quantized subspace (ADQSP)
We show that by varying a single quantization parameter the proposed method can vary between SMPC-type performances and DP-type performances.
Our results show the potential of exploiting traditional distributed signal processing tools for providing cryptographic guarantees.
arXiv Detail & Related papers (2023-12-13T07:52:16Z) - Differentially Private SGD Without Clipping Bias: An Error-Feedback Approach [62.000948039914135]
Using Differentially Private Gradient Descent with Gradient Clipping (DPSGD-GC) to ensure Differential Privacy (DP) comes at the cost of model performance degradation.
We propose a new error-feedback (EF) DP algorithm as an alternative to DPSGD-GC.
We establish an algorithm-specific DP analysis for our proposed algorithm, providing privacy guarantees based on R'enyi DP.
arXiv Detail & Related papers (2023-11-24T17:56:44Z) - Differentially Private Multi-Site Treatment Effect Estimation [28.13660104055298]
Most patient data remains in silo in separate hospitals, preventing the design of data-driven healthcare AI systems.
We look at estimating the average treatment effect (ATE), an important task in causal inference for healthcare applications.
We address this through a class of per-site estimation algorithms that reports the ATE estimate and its variance as a quality measure.
arXiv Detail & Related papers (2023-10-10T01:21:01Z) - Causal Inference with Differentially Private (Clustered) Outcomes [16.166525280886578]
Estimating causal effects from randomized experiments is only feasible if participants agree to reveal their responses.
We suggest a new differential privacy mechanism, Cluster-DP, which leverages any given cluster structure.
We show that, depending on an intuitive measure of cluster quality, we can improve the variance loss while maintaining our privacy guarantees.
arXiv Detail & Related papers (2023-08-02T05:51:57Z) - A Differentially Private Weighted Empirical Risk Minimization Procedure and its Application to Outcome Weighted Learning [4.322221694511603]
Differential privacy (DP) is an appealing framework for addressing data privacy issues.
DP provides mathematically provable bounds on the privacy loss incurred when releasing information from sensitive data.
We propose the first differentially private algorithm for general wERM, with theoretical DP guarantees.
arXiv Detail & Related papers (2023-07-24T21:03:25Z) - B-Learner: Quasi-Oracle Bounds on Heterogeneous Causal Effects Under
Hidden Confounding [51.74479522965712]
We propose a meta-learner called the B-Learner, which can efficiently learn sharp bounds on the CATE function under limits on hidden confounding.
We prove its estimates are valid, sharp, efficient, and have a quasi-oracle property with respect to the constituent estimators under more general conditions than existing methods.
arXiv Detail & Related papers (2023-04-20T18:07:19Z) - Partial Identification with Noisy Covariates: A Robust Optimization
Approach [94.10051154390237]
Causal inference from observational datasets often relies on measuring and adjusting for covariates.
We show that this robust optimization approach can extend a wide range of causal adjustment methods to perform partial identification.
Across synthetic and real datasets, we find that this approach provides ATE bounds with a higher coverage probability than existing methods.
arXiv Detail & Related papers (2022-02-22T04:24:26Z) - Differentially Private Federated Learning with Laplacian Smoothing [72.85272874099644]
Federated learning aims to protect data privacy by collaboratively learning a model without sharing private data among users.
An adversary may still be able to infer the private training data by attacking the released model.
Differential privacy provides a statistical protection against such attacks at the price of significantly degrading the accuracy or utility of the trained models.
arXiv Detail & Related papers (2020-05-01T04:28:38Z) - Machine learning for causal inference: on the use of cross-fit
estimators [77.34726150561087]
Doubly-robust cross-fit estimators have been proposed to yield better statistical properties.
We conducted a simulation study to assess the performance of several estimators for the average causal effect (ACE)
When used with machine learning, the doubly-robust cross-fit estimators substantially outperformed all of the other estimators in terms of bias, variance, and confidence interval coverage.
arXiv Detail & Related papers (2020-04-21T23:09:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.