A Differentially Private Weighted Empirical Risk Minimization Procedure
and its Application to Outcome Weighted Learning
- URL: http://arxiv.org/abs/2307.13127v1
- Date: Mon, 24 Jul 2023 21:03:25 GMT
- Title: A Differentially Private Weighted Empirical Risk Minimization Procedure
and its Application to Outcome Weighted Learning
- Authors: Spencer Giddens, Yiwang Zhou, Kevin R. Krull, Tara M. Brinkman, Peter
X.K. Song, Fang Liu
- Abstract summary: We propose the first differentially private wERM algorithm, backed by a rigorous theoretical proof of its DP guarantees.
We evaluate the performance of the DP-wERM application to weighted learning (OWL) in a simulation study and in a real clinical trial of melatonin for sleep health.
- Score: 5.025486694392673
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: It is commonplace to use data containing personal information to build
predictive models in the framework of empirical risk minimization (ERM). While
these models can be highly accurate in prediction, results obtained from these
models with the use of sensitive data may be susceptible to privacy attacks.
Differential privacy (DP) is an appealing framework for addressing such data
privacy issues by providing mathematically provable bounds on the privacy loss
incurred when releasing information from sensitive data. Previous work has
primarily concentrated on applying DP to unweighted ERM. We consider an
important generalization to weighted ERM (wERM). In wERM, each individual's
contribution to the objective function can be assigned varying weights. In this
context, we propose the first differentially private wERM algorithm, backed by
a rigorous theoretical proof of its DP guarantees under mild regularity
conditions. Extending the existing DP-ERM procedures to wERM paves a path to
deriving privacy-preserving learning methods for individualized treatment
rules, including the popular outcome weighted learning (OWL). We evaluate the
performance of the DP-wERM application to OWL in a simulation study and in a
real clinical trial of melatonin for sleep health. All empirical results
demonstrate the viability of training OWL models via wERM with DP guarantees
while maintaining sufficiently useful model performance. Therefore, we
recommend practitioners consider implementing the proposed privacy-preserving
OWL procedure in real-world scenarios involving sensitive data.
Related papers
- LLM-based Privacy Data Augmentation Guided by Knowledge Distillation
with a Distribution Tutor for Medical Text Classification [67.92145284679623]
We propose a DP-based tutor that models the noised private distribution and controls samples' generation with a low privacy cost.
We theoretically analyze our model's privacy protection and empirically verify our model.
arXiv Detail & Related papers (2024-02-26T11:52:55Z) - Conditional Density Estimations from Privacy-Protected Data [0.0]
We propose simulation-based inference methods from privacy-protected datasets.
We illustrate our methods on discrete time-series data under an infectious disease model and with ordinary linear regression models.
arXiv Detail & Related papers (2023-10-19T14:34:17Z) - MedDiffusion: Boosting Health Risk Prediction via Diffusion-based Data
Augmentation [58.93221876843639]
This paper introduces a novel, end-to-end diffusion-based risk prediction model, named MedDiffusion.
It enhances risk prediction performance by creating synthetic patient data during training to enlarge sample space.
It discerns hidden relationships between patient visits using a step-wise attention mechanism, enabling the model to automatically retain the most vital information for generating high-quality data.
arXiv Detail & Related papers (2023-10-04T01:36:30Z) - Differential Privacy for Adaptive Weight Aggregation in Federated Tumor
Segmentation [0.16746114653388383]
Federated Learning (FL) is a distributed machine learning approach that safeguards privacy by creating an impartial global model while respecting the privacy of individual client data.
We present a differential privacy (DP) federated deep learning framework in medical image segmentation.
We extend our similarity weight aggregation (SimAgg) method to DP-SimAgg algorithm, a differentially private similarity-weighted aggregation algorithm for brain tumor segmentation.
arXiv Detail & Related papers (2023-08-01T21:59:22Z) - Large Language Models for Healthcare Data Augmentation: An Example on
Patient-Trial Matching [49.78442796596806]
We propose an innovative privacy-aware data augmentation approach for patient-trial matching (LLM-PTM)
Our experiments demonstrate a 7.32% average improvement in performance using the proposed LLM-PTM method, and the generalizability to new data is improved by 12.12%.
arXiv Detail & Related papers (2023-03-24T03:14:00Z) - Policy Optimization for Personalized Interventions in Behavioral Health [8.10897203067601]
Behavioral health interventions, delivered through digital platforms, have the potential to significantly improve health outcomes.
We study the problem of optimizing personalized interventions for patients to maximize a long-term outcome.
We present a new approach for this problem that we dub DecompPI, which decomposes the state space for a system of patients to the individual level.
arXiv Detail & Related papers (2023-03-21T21:42:03Z) - No Free Lunch in "Privacy for Free: How does Dataset Condensation Help
Privacy" [75.98836424725437]
New methods designed to preserve data privacy require careful scrutiny.
Failure to preserve privacy is hard to detect, and yet can lead to catastrophic results when a system implementing a privacy-preserving'' method is attacked.
arXiv Detail & Related papers (2022-09-29T17:50:23Z) - Differentially Private Estimation of Heterogeneous Causal Effects [9.355532300027727]
We introduce a general meta-algorithm for estimating conditional average treatment effects (CATE) with differential privacy guarantees.
Our meta-algorithm can work with simple, single-stage CATE estimators such as S-learner and more complex multi-stage estimators such as DR and R-learner.
arXiv Detail & Related papers (2022-02-22T17:21:18Z) - DP-UTIL: Comprehensive Utility Analysis of Differential Privacy in
Machine Learning [3.822543555265593]
Differential Privacy (DP) has emerged as a rigorous formalism to reason about privacy leakage.
In machine learning (ML), DP has been employed to limit/disclosure of training examples.
For deep neural networks, gradient perturbation results in lowest privacy leakage.
arXiv Detail & Related papers (2021-12-24T08:40:28Z) - Differentially private federated deep learning for multi-site medical
image segmentation [56.30543374146002]
Collaborative machine learning techniques such as federated learning (FL) enable the training of models on effectively larger datasets without data transfer.
Recent initiatives have demonstrated that segmentation models trained with FL can achieve performance similar to locally trained models.
However, FL is not a fully privacy-preserving technique and privacy-centred attacks can disclose confidential patient data.
arXiv Detail & Related papers (2021-07-06T12:57:32Z) - Privacy-preserving medical image analysis [53.4844489668116]
We present PriMIA, a software framework designed for privacy-preserving machine learning (PPML) in medical imaging.
We show significantly better classification performance of a securely aggregated federated learning model compared to human experts on unseen datasets.
We empirically evaluate the framework's security against a gradient-based model inversion attack.
arXiv Detail & Related papers (2020-12-10T13:56:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.