Amnesiac Machine Learning
- URL: http://arxiv.org/abs/2010.10981v1
- Date: Wed, 21 Oct 2020 13:14:17 GMT
- Title: Amnesiac Machine Learning
- Authors: Laura Graves, Vineel Nagisetty, Vijay Ganesh
- Abstract summary: Recently enacted General Data Protection Regulation affects any data holder that has data on European Union residents.
Models are vulnerable to information leaking attacks such as model inversion attacks.
We present two data removal methods, namely Unlearning and Amnesiac Unlearning, that enable model owners to protect themselves against such attacks while being compliant with regulations.
- Score: 15.680008735220785
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The Right to be Forgotten is part of the recently enacted General Data
Protection Regulation (GDPR) law that affects any data holder that has data on
European Union residents. It gives EU residents the ability to request deletion
of their personal data, including training records used to train machine
learning models. Unfortunately, Deep Neural Network models are vulnerable to
information leaking attacks such as model inversion attacks which extract class
information from a trained model and membership inference attacks which
determine the presence of an example in a model's training data. If a malicious
party can mount an attack and learn private information that was meant to be
removed, then it implies that the model owner has not properly protected their
user's rights and their models may not be compliant with the GDPR law. In this
paper, we present two efficient methods that address this question of how a
model owner or data holder may delete personal data from models in such a way
that they may not be vulnerable to model inversion and membership inference
attacks while maintaining model efficacy. We start by presenting a real-world
threat model that shows that simply removing training data is insufficient to
protect users. We follow that up with two data removal methods, namely
Unlearning and Amnesiac Unlearning, that enable model owners to protect
themselves against such attacks while being compliant with regulations. We
provide extensive empirical analysis that show that these methods are indeed
efficient, safe to apply, effectively remove learned information about
sensitive data from trained models while maintaining model efficacy.
Related papers
- Membership Inference Attacks Cannot Prove that a Model Was Trained On Your Data [27.18781946018255]
Training data proofs play a key role in recent lawsuits against foundation models trained on web-scale data.
Many prior works suggest to instantiate training data proofs using membership inference attacks.
We show that data extraction attacks and membership inference on special canary data can be used to create sound training data proofs.
arXiv Detail & Related papers (2024-09-29T21:49:32Z) - Releasing Malevolence from Benevolence: The Menace of Benign Data on Machine Unlearning [28.35038726318893]
Machine learning models trained on vast amounts of real or synthetic data often achieve outstanding predictive performance across various domains.
To address privacy concerns, machine unlearning has been proposed to erase specific data samples from models.
We introduce the Unlearning Usability Attack to distill data distribution information into a small set of benign data.
arXiv Detail & Related papers (2024-07-06T15:42:28Z) - Reconstruction Attacks on Machine Unlearning: Simple Models are Vulnerable [30.22146634953896]
We show how to mount a near-perfect attack on the deleted data point from linear regression models.
Our work highlights that privacy risk is significant even for extremely simple model classes when individuals can request deletion of their data from the model.
arXiv Detail & Related papers (2024-05-30T17:27:44Z) - Beyond Labeling Oracles: What does it mean to steal ML models? [52.63413852460003]
Model extraction attacks are designed to steal trained models with only query access.
We investigate factors influencing the success of model extraction attacks.
Our findings urge the community to redefine the adversarial goals of ME attacks.
arXiv Detail & Related papers (2023-10-03T11:10:21Z) - AI Model Disgorgement: Methods and Choices [127.54319351058167]
We introduce a taxonomy of possible disgorgement methods that are applicable to modern machine learning systems.
We investigate the meaning of "removing the effects" of data in the trained model in a way that does not require retraining from scratch.
arXiv Detail & Related papers (2023-04-07T08:50:18Z) - CANIFE: Crafting Canaries for Empirical Privacy Measurement in Federated
Learning [77.27443885999404]
Federated Learning (FL) is a setting for training machine learning models in distributed environments.
We propose a novel method, CANIFE, that uses carefully crafted samples by a strong adversary to evaluate the empirical privacy of a training round.
arXiv Detail & Related papers (2022-10-06T13:30:16Z) - Certified Data Removal in Sum-Product Networks [78.27542864367821]
Deleting the collected data is often insufficient to guarantee data privacy.
UnlearnSPN is an algorithm that removes the influence of single data points from a trained sum-product network.
arXiv Detail & Related papers (2022-10-04T08:22:37Z) - MOVE: Effective and Harmless Ownership Verification via Embedded
External Features [109.19238806106426]
We propose an effective and harmless model ownership verification (MOVE) to defend against different types of model stealing simultaneously.
We conduct the ownership verification by verifying whether a suspicious model contains the knowledge of defender-specified external features.
In particular, we develop our MOVE method under both white-box and black-box settings to provide comprehensive model protection.
arXiv Detail & Related papers (2022-08-04T02:22:29Z) - Machine unlearning via GAN [2.406359246841227]
Machine learning models, especially deep models, may unintentionally remember information about their training data.
We present a GAN-based algorithm to delete data in deep models, which significantly improves deleting speed compared to retraining from scratch.
arXiv Detail & Related papers (2021-11-22T05:28:57Z) - Certified Data Removal from Machine Learning Models [79.91502073022602]
Good data stewardship requires removal of data at the request of the data's owner.
This raises the question if and how a trained machine-learning model, which implicitly stores information about its training data, should be affected by such a removal request.
We study this problem by defining certified removal: a very strong theoretical guarantee that a model from which data is removed cannot be distinguished from a model that never observed the data to begin with.
arXiv Detail & Related papers (2019-11-08T03:57:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.