Collaborative Drug Discovery: Inference-level Data Protection
Perspective
- URL: http://arxiv.org/abs/2205.06506v1
- Date: Fri, 13 May 2022 08:30:50 GMT
- Title: Collaborative Drug Discovery: Inference-level Data Protection
Perspective
- Authors: Balazs Pejo, Mina Remeli, Adam Arany, Mathieu Galtier, Gergely Acs
- Abstract summary: Pharmaceutical industry can better leverage its data assets to virtualize drug discovery through a collaborative machine learning platform.
There are non-negligible risks stemming from the unintended leakage of participants' training data.
This paper describes a privacy risk assessment for collaborative modeling in the preclinical phase of drug discovery to accelerate the selection of promising drug candidates.
- Score: 2.624902795082451
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Pharmaceutical industry can better leverage its data assets to virtualize
drug discovery through a collaborative machine learning platform. On the other
hand, there are non-negligible risks stemming from the unintended leakage of
participants' training data, hence, it is essential for such a platform to be
secure and privacy-preserving. This paper describes a privacy risk assessment
for collaborative modeling in the preclinical phase of drug discovery to
accelerate the selection of promising drug candidates. After a short taxonomy
of state-of-the-art inference attacks we adopt and customize several to the
underlying scenario. Finally we describe and experiments with a handful of
relevant privacy protection techniques to mitigate such attacks.
Related papers
- In-depth Analysis of Privacy Threats in Federated Learning for Medical Data [2.6986500640871482]
Federated learning is emerging as a promising machine learning technique in the medical field for analyzing medical images.
Recent studies have revealed that the default settings of federated learning may inadvertently expose private training data to privacy attacks.
We make three original contributions to privacy risk analysis and mitigation in federated learning for medical data.
arXiv Detail & Related papers (2024-09-27T16:45:35Z) - Privacy-Preserving Collaborative Genomic Research: A Real-Life Deployment and Vision [2.7968600664591983]
This paper presents a privacy-preserving framework for genomic research, developed in collaboration with Lynx.MD.
The framework addresses critical cybersecurity and privacy challenges, enabling the privacy-preserving sharing and analysis of genomic data.
Implementing the framework within Lynx.MD involves encoding genomic data into binary formats and applying noise through controlled perturbation techniques.
arXiv Detail & Related papers (2024-07-12T05:43:13Z) - Collection, usage and privacy of mobility data in the enterprise and public administrations [55.2480439325792]
Security measures such as anonymization are needed to protect individuals' privacy.
Within our study, we conducted expert interviews to gain insights into practices in the field.
We survey privacy-enhancing methods in use, which generally do not comply with state-of-the-art standards of differential privacy.
arXiv Detail & Related papers (2024-07-04T08:29:27Z) - A Unified View of Differentially Private Deep Generative Modeling [60.72161965018005]
Data with privacy concerns comes with stringent regulations that frequently prohibited data access and data sharing.
Overcoming these obstacles is key for technological progress in many real-world application scenarios that involve privacy sensitive data.
Differentially private (DP) data publishing provides a compelling solution, where only a sanitized form of the data is publicly released.
arXiv Detail & Related papers (2023-09-27T14:38:16Z) - On the Privacy Risks of Algorithmic Recourse [17.33484111779023]
We make the first attempt at investigating if and how an adversary can leverage recourses to infer private information about the underlying model's training data.
Our work establishes unintended privacy leakage as an important risk in the widespread adoption of recourse methods.
arXiv Detail & Related papers (2022-11-10T09:04:24Z) - SSM-DTA: Breaking the Barriers of Data Scarcity in Drug-Target Affinity
Prediction [127.43571146741984]
Drug-Target Affinity (DTA) is of vital importance in early-stage drug discovery.
wet experiments remain the most reliable method, but they are time-consuming and resource-intensive.
Existing methods have primarily focused on developing techniques based on the available DTA data, without adequately addressing the data scarcity issue.
We present the SSM-DTA framework, which incorporates three simple yet highly effective strategies.
arXiv Detail & Related papers (2022-06-20T14:53:25Z) - PeQES: A Platform for Privacy-enhanced Quantitative Empirical Studies [6.782635275179198]
We establish a novel, privacy-enhanced workflow for pre-registered studies.
We also introduce PeQES, a corresponding platform that technically enforces the appropriate execution.
PeQES is the first platform to enable privacy-enhanced studies, to ensure the integrity of study protocols, and to safeguard the confidentiality of participants' data at the same time.
arXiv Detail & Related papers (2021-03-09T16:46:25Z) - Defending Medical Image Diagnostics against Privacy Attacks using
Generative Methods [10.504951891644474]
We develop and evaluate a privacy defense protocol based on using a generative adversarial network (GAN)
We validate the proposed method on retinal diagnostics AI used for diabetic retinopathy that bears the risk of possibly leaking private information.
arXiv Detail & Related papers (2021-03-04T15:02:57Z) - Privacy-preserving medical image analysis [53.4844489668116]
We present PriMIA, a software framework designed for privacy-preserving machine learning (PPML) in medical imaging.
We show significantly better classification performance of a securely aggregated federated learning model compared to human experts on unseen datasets.
We empirically evaluate the framework's security against a gradient-based model inversion attack.
arXiv Detail & Related papers (2020-12-10T13:56:00Z) - Epidemic mitigation by statistical inference from contact tracing data [61.04165571425021]
We develop Bayesian inference methods to estimate the risk that an individual is infected.
We propose to use probabilistic risk estimation in order to optimize testing and quarantining strategies for the control of an epidemic.
Our approaches translate into fully distributed algorithms that only require communication between individuals who have recently been in contact.
arXiv Detail & Related papers (2020-09-20T12:24:45Z) - COVI White Paper [67.04578448931741]
Contact tracing is an essential tool to change the course of the Covid-19 pandemic.
We present an overview of the rationale, design, ethical considerations and privacy strategy of COVI,' a Covid-19 public peer-to-peer contact tracing and risk awareness mobile application developed in Canada.
arXiv Detail & Related papers (2020-05-18T07:40:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.