Machine Unlearning: Taxonomy, Metrics, Applications, Challenges, and
Prospects
- URL: http://arxiv.org/abs/2403.08254v1
- Date: Wed, 13 Mar 2024 05:11:24 GMT
- Title: Machine Unlearning: Taxonomy, Metrics, Applications, Challenges, and
Prospects
- Authors: Na Li, Chunyi Zhou, Yansong Gao, Hui Chen, Anmin Fu, Zhi Zhang, and Yu
Shui
- Abstract summary: Data users have been endowed with the right to be forgotten of their data.
In the course of machine learning (ML), the forgotten right requires a model provider to delete user data.
Machine unlearning emerges to address this, which has garnered ever-increasing attention from both industry and academia.
- Score: 17.502158848870426
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Personal digital data is a critical asset, and governments worldwide have
enforced laws and regulations to protect data privacy. Data users have been
endowed with the right to be forgotten of their data. In the course of machine
learning (ML), the forgotten right requires a model provider to delete user
data and its subsequent impact on ML models upon user requests. Machine
unlearning emerges to address this, which has garnered ever-increasing
attention from both industry and academia. While the area has developed
rapidly, there is a lack of comprehensive surveys to capture the latest
advancements. Recognizing this shortage, we conduct an extensive exploration to
map the landscape of machine unlearning including the (fine-grained) taxonomy
of unlearning algorithms under centralized and distributed settings, debate on
approximate unlearning, verification and evaluation metrics, challenges and
solutions for unlearning under different applications, as well as attacks
targeting machine unlearning. The survey concludes by outlining potential
directions for future research, hoping to serve as a guide for interested
scholars.
Related papers
- MUSE: Machine Unlearning Six-Way Evaluation for Language Models [109.76505405962783]
Language models (LMs) are trained on vast amounts of text data, which may include private and copyrighted content.
We propose MUSE, a comprehensive machine unlearning evaluation benchmark.
We benchmark how effectively eight popular unlearning algorithms can unlearn Harry Potter books and news articles.
arXiv Detail & Related papers (2024-07-08T23:47:29Z) - Machine Unlearning for Traditional Models and Large Language Models: A Short Survey [11.539080008361662]
Machine unlearning aims to delete data and reduce its impact on models according to user requests.
This paper categorizes and investigates unlearning on both traditional models and Large Language Models (LLMs)
arXiv Detail & Related papers (2024-04-01T16:08:18Z) - The Frontier of Data Erasure: Machine Unlearning for Large Language Models [56.26002631481726]
Large Language Models (LLMs) are foundational to AI advancements.
LLMs pose risks by potentially memorizing and disseminating sensitive, biased, or copyrighted information.
Machine unlearning emerges as a cutting-edge solution to mitigate these concerns.
arXiv Detail & Related papers (2024-03-23T09:26:15Z) - Threats, Attacks, and Defenses in Machine Unlearning: A Survey [15.05662521329346]
Machine Unlearning (MU) has gained considerable attention recently for its potential to achieve Safe AI.
This survey aims to fill the gap between the extensive number of studies on threats, attacks, and defenses in machine unlearning.
arXiv Detail & Related papers (2024-03-20T15:40:18Z) - On Responsible Machine Learning Datasets with Fairness, Privacy, and
Regulatory Norms [58.93352076927003]
There have been severe concerns over the trustworthiness of AI technologies.
Machine and deep learning algorithms depend heavily on the data used during their development.
We propose a framework to evaluate the datasets through a responsible rubric.
arXiv Detail & Related papers (2023-10-24T14:01:53Z) - Machine Unlearning: A Survey [56.79152190680552]
A special need has arisen where, due to privacy, usability, and/or the right to be forgotten, information about some specific samples needs to be removed from a model, called machine unlearning.
This emerging technology has drawn significant interest from both academics and industry due to its innovation and practicality.
No study has analyzed this complex topic or compared the feasibility of existing unlearning solutions in different kinds of scenarios.
The survey concludes by highlighting some of the outstanding issues with unlearning techniques, along with some feasible directions for new research opportunities.
arXiv Detail & Related papers (2023-06-06T10:18:36Z) - Exploring the Landscape of Machine Unlearning: A Comprehensive Survey
and Taxonomy [17.535417441295074]
Machine unlearning (MU) is gaining increasing attention due to the need to remove or modify predictions made by machine learning (ML) models.
This paper presents a comprehensive survey of MU, covering current state-of-the-art techniques and approaches.
The paper also highlights the challenges that need to be addressed, including attack sophistication, standardization, transferability, interpretability, and resource constraints.
arXiv Detail & Related papers (2023-05-10T12:02:18Z) - Privacy Adhering Machine Un-learning in NLP [66.17039929803933]
In real world industry use Machine Learning to build models on user data.
Such mandates require effort both in terms of data as well as model retraining.
continuous removal of data and model retraining steps do not scale.
We propose textitMachine Unlearning to tackle this challenge.
arXiv Detail & Related papers (2022-12-19T16:06:45Z) - A Survey of Machine Unlearning [43.272767023563254]
Recent regulations require that private information about a user can be removed from computer systems in general and from ML models in particular upon request.
This phenomenon calls for a new paradigm, namely machine unlearning, to make ML models forget about particular data.
We seek to provide a thorough investigation of machine unlearning in its definitions, scenarios, mechanisms, and applications.
arXiv Detail & Related papers (2022-09-06T08:51:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.