Zero-Shot Machine Unlearning
- URL: http://arxiv.org/abs/2201.05629v3
- Date: Wed, 31 May 2023 17:13:20 GMT
- Title: Zero-Shot Machine Unlearning
- Authors: Vikram S Chundawat, Ayush K Tarun, Murari Mandal, Mohan Kankanhalli
- Abstract summary: Modern privacy regulations grant citizens the right to be forgotten by products, services and companies.
No data related to the training process or training samples may be accessible for the unlearning purpose.
We propose two novel solutions for zero-shot machine unlearning based on (a) error minimizing-maximizing noise and (b) gated knowledge transfer.
- Score: 6.884272840652062
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Modern privacy regulations grant citizens the right to be forgotten by
products, services and companies. In case of machine learning (ML)
applications, this necessitates deletion of data not only from storage archives
but also from ML models. Due to an increasing need for regulatory compliance
required for ML applications, machine unlearning is becoming an emerging
research problem. The right to be forgotten requests come in the form of
removal of a certain set or class of data from the already trained ML model.
Practical considerations preclude retraining of the model from scratch after
discarding the deleted data. The few existing studies use either the whole
training data, or a subset of training data, or some metadata stored during
training to update the model weights for unlearning. However, in many cases, no
data related to the training process or training samples may be accessible for
the unlearning purpose. We therefore ask the question: is it possible to
achieve unlearning with zero training samples? In this paper, we introduce the
novel problem of zero-shot machine unlearning that caters for the extreme but
practical scenario where zero original data samples are available for use. We
then propose two novel solutions for zero-shot machine unlearning based on (a)
error minimizing-maximizing noise and (b) gated knowledge transfer. These
methods remove the information of the forget data from the model while
maintaining the model efficacy on the retain data. The zero-shot approach
offers good protection against the model inversion attacks and membership
inference attacks. We introduce a new evaluation metric, Anamnesis Index (AIN)
to effectively measure the quality of the unlearning method. The experiments
show promising results for unlearning in deep learning models on benchmark
vision data-sets. The source code is available here:
https://github.com/ayu987/zero-shot-unlearning
Related papers
- Attribute-to-Delete: Machine Unlearning via Datamodel Matching [65.13151619119782]
Machine unlearning -- efficiently removing a small "forget set" training data on a pre-divertrained machine learning model -- has recently attracted interest.
Recent research shows that machine unlearning techniques do not hold up in such a challenging setting.
arXiv Detail & Related papers (2024-10-30T17:20:10Z) - Learn to Unlearn for Deep Neural Networks: Minimizing Unlearning
Interference with Gradient Projection [56.292071534857946]
Recent data-privacy laws have sparked interest in machine unlearning.
Challenge is to discard information about the forget'' data without altering knowledge about remaining dataset.
We adopt a projected-gradient based learning method, named as Projected-Gradient Unlearning (PGU)
We provide empirically evidence to demonstrate that our unlearning method can produce models that behave similar to models retrained from scratch across various metrics even when the training dataset is no longer accessible.
arXiv Detail & Related papers (2023-12-07T07:17:24Z) - Learn to Unlearn: A Survey on Machine Unlearning [29.077334665555316]
This article presents a review of recent machine unlearning techniques, verification mechanisms, and potential attacks.
We highlight emerging challenges and prospective research directions.
We aim for this paper to provide valuable resources for integrating privacy, equity, andresilience into ML systems.
arXiv Detail & Related papers (2023-05-12T14:28:02Z) - AI Model Disgorgement: Methods and Choices [127.54319351058167]
We introduce a taxonomy of possible disgorgement methods that are applicable to modern machine learning systems.
We investigate the meaning of "removing the effects" of data in the trained model in a way that does not require retraining from scratch.
arXiv Detail & Related papers (2023-04-07T08:50:18Z) - Deep Regression Unlearning [6.884272840652062]
We introduce deep regression unlearning methods that generalize well and are robust to privacy attacks.
We conduct regression unlearning experiments for computer vision, natural language processing and forecasting applications.
arXiv Detail & Related papers (2022-10-15T05:00:20Z) - A Survey of Machine Unlearning [56.017968863854186]
Recent regulations now require that, on request, private information about a user must be removed from computer systems.
ML models often remember' the old data.
Recent works on machine unlearning have not been able to completely solve the problem.
arXiv Detail & Related papers (2022-09-06T08:51:53Z) - Machine Unlearning of Features and Labels [72.81914952849334]
We propose first scenarios for unlearning and labels in machine learning models.
Our approach builds on the concept of influence functions and realizes unlearning through closed-form updates of model parameters.
arXiv Detail & Related papers (2021-08-26T04:42:24Z) - SSSE: Efficiently Erasing Samples from Trained Machine Learning Models [103.43466657962242]
We propose an efficient and effective algorithm, SSSE, for samples erasure.
In certain cases SSSE can erase samples almost as well as the optimal, yet impractical, gold standard of training a new model from scratch with only the permitted data.
arXiv Detail & Related papers (2021-07-08T14:17:24Z) - Certifiable Machine Unlearning for Linear Models [1.484852576248587]
Machine unlearning is the task of updating machine learning (ML) models after a subset of the training data they were trained on is deleted.
We present an experimental study of the three state-of-the-art approximate unlearning methods for linear models.
arXiv Detail & Related papers (2021-06-29T05:05:58Z) - Supervised Machine Learning with Plausible Deniability [1.685485565763117]
We study the question of how well machine learning (ML) models trained on a certain data set provide privacy for the training data.
We show that one can take a set of purely random training data, and from this define a suitable learning rule'' that will produce a ML model that is exactly $f$.
arXiv Detail & Related papers (2021-06-08T11:54:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.