Tight Bounds for Machine Unlearning via Differential Privacy
- URL: http://arxiv.org/abs/2309.00886v1
- Date: Sat, 2 Sep 2023 09:55:29 GMT
- Title: Tight Bounds for Machine Unlearning via Differential Privacy
- Authors: Yiyang Huang, Cl\'ement L. Canonne
- Abstract summary: We consider the so-called "right to be forgotten" by requiring that a trained model should be able to "unlearn" a number of points from the training data.
We obtain tight bounds on the deletion capacity achievable by DP-based machine unlearning algorithms.
- Score: 0.7252027234425334
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We consider the formulation of "machine unlearning" of Sekhari, Acharya,
Kamath, and Suresh (NeurIPS 2021), which formalizes the so-called "right to be
forgotten" by requiring that a trained model, upon request, should be able to
"unlearn" a number of points from the training data, as if they had never been
included in the first place. Sekhari et al. established some positive and
negative results about the number of data points that can be successfully
unlearnt by a trained model without impacting the model's accuracy (the
"deletion capacity"), showing that machine unlearning could be achieved by
using differentially private (DP) algorithms. However, their results left open
a gap between upper and lower bounds on the deletion capacity of these
algorithms: our work fully closes this gap, obtaining tight bounds on the
deletion capacity achievable by DP-based machine unlearning algorithms.
Related papers
- RESTOR: Knowledge Recovery through Machine Unlearning [71.75834077528305]
Large language models trained on web-scale corpora can memorize undesirable datapoints.
Many machine unlearning methods have been proposed that aim to 'erase' these datapoints from trained models.
We propose the RESTOR framework for machine unlearning based on the following dimensions.
arXiv Detail & Related papers (2024-10-31T20:54:35Z) - Attribute-to-Delete: Machine Unlearning via Datamodel Matching [65.13151619119782]
Machine unlearning -- efficiently removing a small "forget set" training data on a pre-divertrained machine learning model -- has recently attracted interest.
Recent research shows that machine unlearning techniques do not hold up in such a challenging setting.
arXiv Detail & Related papers (2024-10-30T17:20:10Z) - How to unlearn a learned Machine Learning model ? [0.0]
I will present an elegant algorithm for unlearning a machine learning model and visualize its abilities.
I will elucidate the underlying mathematical theory and establish specific metrics to evaluate both the unlearned model's performance on desired data and its level of ignorance regarding unwanted data.
arXiv Detail & Related papers (2024-10-13T17:38:09Z) - Rewind-to-Delete: Certified Machine Unlearning for Nonconvex Functions [11.955062839855334]
Machine unlearning algorithms aim to efficiently data from a model without it from scratch, in order to enforce data privacy, remove corrupted or outdated data, or respect a user's right to forgotten"
Our algorithm is black-box, in that it be directly applied to models with vanilla gradient descent with no prior consideration of unlearning.
arXiv Detail & Related papers (2024-09-15T15:58:08Z) - MUSE: Machine Unlearning Six-Way Evaluation for Language Models [109.76505405962783]
Language models (LMs) are trained on vast amounts of text data, which may include private and copyrighted content.
We propose MUSE, a comprehensive machine unlearning evaluation benchmark.
We benchmark how effectively eight popular unlearning algorithms can unlearn Harry Potter books and news articles.
arXiv Detail & Related papers (2024-07-08T23:47:29Z) - Learn to Unlearn for Deep Neural Networks: Minimizing Unlearning
Interference with Gradient Projection [56.292071534857946]
Recent data-privacy laws have sparked interest in machine unlearning.
Challenge is to discard information about the forget'' data without altering knowledge about remaining dataset.
We adopt a projected-gradient based learning method, named as Projected-Gradient Unlearning (PGU)
We provide empirically evidence to demonstrate that our unlearning method can produce models that behave similar to models retrained from scratch across various metrics even when the training dataset is no longer accessible.
arXiv Detail & Related papers (2023-12-07T07:17:24Z) - Forget Unlearning: Towards True Data-Deletion in Machine Learning [18.656957502454592]
We show that unlearning is not equivalent to data deletion and does not guarantee the "right to be forgotten"
We propose an accurate, computationally efficient, and secure data-deletion machine learning algorithm in the online setting.
arXiv Detail & Related papers (2022-10-17T10:06:11Z) - Evaluating Machine Unlearning via Epistemic Uncertainty [78.27542864367821]
This work presents an evaluation of Machine Unlearning algorithms based on uncertainty.
This is the first definition of a general evaluation of our best knowledge.
arXiv Detail & Related papers (2022-08-23T09:37:31Z) - On the Necessity of Auditable Algorithmic Definitions for Machine
Unlearning [13.149070833843133]
Machine unlearning, i.e. having a model forget about some of its training data, has become increasingly important as privacy legislation promotes variants of the right-to-be-forgotten.
We first show that the definition that underlies approximate unlearning, which seeks to prove the approximately unlearned model is close to an exactly retrained model, is incorrect because one can obtain the same model using different datasets.
We then turn to exact unlearning approaches and ask how to verify their claims of unlearning.
arXiv Detail & Related papers (2021-10-22T16:16:56Z) - Machine Unlearning of Features and Labels [72.81914952849334]
We propose first scenarios for unlearning and labels in machine learning models.
Our approach builds on the concept of influence functions and realizes unlearning through closed-form updates of model parameters.
arXiv Detail & Related papers (2021-08-26T04:42:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.