Exploring the Landscape of Machine Unlearning: A Comprehensive Survey
and Taxonomy
- URL: http://arxiv.org/abs/2305.06360v6
- Date: Thu, 1 Feb 2024 01:07:22 GMT
- Title: Exploring the Landscape of Machine Unlearning: A Comprehensive Survey
and Taxonomy
- Authors: Thanveer Shaik, Xiaohui Tao, Haoran Xie, Lin Li, Xiaofeng Zhu, and
Qing Li
- Abstract summary: Machine unlearning (MU) is gaining increasing attention due to the need to remove or modify predictions made by machine learning (ML) models.
This paper presents a comprehensive survey of MU, covering current state-of-the-art techniques and approaches.
The paper also highlights the challenges that need to be addressed, including attack sophistication, standardization, transferability, interpretability, and resource constraints.
- Score: 17.535417441295074
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Machine unlearning (MU) is gaining increasing attention due to the need to
remove or modify predictions made by machine learning (ML) models. While
training models have become more efficient and accurate, the importance of
unlearning previously learned information has become increasingly significant
in fields such as privacy, security, and fairness. This paper presents a
comprehensive survey of MU, covering current state-of-the-art techniques and
approaches, including data deletion, perturbation, and model updates. In
addition, commonly used metrics and datasets are also presented. The paper also
highlights the challenges that need to be addressed, including attack
sophistication, standardization, transferability, interpretability, training
data, and resource constraints. The contributions of this paper include
discussions about the potential benefits of MU and its future directions.
Additionally, the paper emphasizes the need for researchers and practitioners
to continue exploring and refining unlearning techniques to ensure that ML
models can adapt to changing circumstances while maintaining user trust. The
importance of unlearning is further highlighted in making Artificial
Intelligence (AI) more trustworthy and transparent, especially with the
increasing importance of AI in various domains that involve large amounts of
personal user data.
Related papers
- Federated Learning driven Large Language Models for Swarm Intelligence: A Survey [2.769238399659845]
Federated learning (FL) offers a compelling framework for training large language models (LLMs)
We focus on machine unlearning, a crucial aspect for complying with privacy regulations like the Right to be Forgotten.
We explore various strategies that enable effective unlearning, such as perturbation techniques, model decomposition, and incremental learning.
arXiv Detail & Related papers (2024-06-14T08:40:58Z) - Machine Unlearning for Traditional Models and Large Language Models: A Short Survey [11.539080008361662]
Machine unlearning aims to delete data and reduce its impact on models according to user requests.
This paper categorizes and investigates unlearning on both traditional models and Large Language Models (LLMs)
arXiv Detail & Related papers (2024-04-01T16:08:18Z) - The Frontier of Data Erasure: Machine Unlearning for Large Language Models [56.26002631481726]
Large Language Models (LLMs) are foundational to AI advancements.
LLMs pose risks by potentially memorizing and disseminating sensitive, biased, or copyrighted information.
Machine unlearning emerges as a cutting-edge solution to mitigate these concerns.
arXiv Detail & Related papers (2024-03-23T09:26:15Z) - Machine Unlearning: Taxonomy, Metrics, Applications, Challenges, and
Prospects [17.502158848870426]
Data users have been endowed with the right to be forgotten of their data.
In the course of machine learning (ML), the forgotten right requires a model provider to delete user data.
Machine unlearning emerges to address this, which has garnered ever-increasing attention from both industry and academia.
arXiv Detail & Related papers (2024-03-13T05:11:24Z) - Rethinking Machine Unlearning for Large Language Models [85.92660644100582]
We explore machine unlearning in the domain of large language models (LLMs)
This initiative aims to eliminate undesirable data influence (e.g., sensitive or illegal information) and the associated model capabilities.
arXiv Detail & Related papers (2024-02-13T20:51:58Z) - Deep Transfer Learning for Automatic Speech Recognition: Towards Better
Generalization [3.6393183544320236]
Speech recognition has become an important challenge when using deep learning (DL)
It requires large-scale training datasets and high computational and storage resources.
Deep transfer learning (DTL) has been introduced to overcome these issues.
arXiv Detail & Related papers (2023-04-27T21:08:05Z) - Deep Active Learning for Computer Vision: Past and Future [50.19394935978135]
Despite its indispensable role for developing AI models, research on active learning is not as intensive as other research directions.
By addressing data automation challenges and coping with automated machine learning systems, active learning will facilitate democratization of AI technologies.
arXiv Detail & Related papers (2022-11-27T13:07:14Z) - Learnware: Small Models Do Big [69.88234743773113]
The prevailing big model paradigm, which has achieved impressive results in natural language processing and computer vision applications, has not yet addressed those issues, whereas becoming a serious source of carbon emissions.
This article offers an overview of the learnware paradigm, which attempts to enable users not need to build machine learning models from scratch, with the hope of reusing small models to do things even beyond their original purposes.
arXiv Detail & Related papers (2022-10-07T15:55:52Z) - Pre-Trained Models: Past, Present and Future [126.21572378910746]
Large-scale pre-trained models (PTMs) have recently achieved great success and become a milestone in the field of artificial intelligence (AI)
By storing knowledge into huge parameters and fine-tuning on specific tasks, the rich knowledge implicitly encoded in huge parameters can benefit a variety of downstream tasks.
It is now the consensus of the AI community to adopt PTMs as backbone for downstream tasks rather than learning models from scratch.
arXiv Detail & Related papers (2021-06-14T02:40:32Z) - Personalized Education in the AI Era: What to Expect Next? [76.37000521334585]
The objective of personalized learning is to design an effective knowledge acquisition track that matches the learner's strengths and bypasses her weaknesses to meet her desired goal.
In recent years, the boost of artificial intelligence (AI) and machine learning (ML) has unfolded novel perspectives to enhance personalized education.
arXiv Detail & Related papers (2021-01-19T12:23:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.