Contrastive Unlearning: A Contrastive Approach to Machine Unlearning
- URL: http://arxiv.org/abs/2401.10458v1
- Date: Fri, 19 Jan 2024 02:16:30 GMT
- Title: Contrastive Unlearning: A Contrastive Approach to Machine Unlearning
- Authors: Hong kyu Lee, Qiuchen Zhang, Carl Yang, Jian Lou, Li Xiong
- Abstract summary: We propose a contrastive unlearning framework, leveraging the concept of representation learning for more effective unlearning.
We show that contrastive unlearning achieves the best unlearning effects and efficiency with the lowest performance loss compared with the state-of-the-art algorithms.
- Score: 30.38966646250252
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Machine unlearning aims to eliminate the influence of a subset of training
samples (i.e., unlearning samples) from a trained model. Effectively and
efficiently removing the unlearning samples without negatively impacting the
overall model performance is still challenging. In this paper, we propose a
contrastive unlearning framework, leveraging the concept of representation
learning for more effective unlearning. It removes the influence of unlearning
samples by contrasting their embeddings against the remaining samples so that
they are pushed away from their original classes and pulled toward other
classes. By directly optimizing the representation space, it effectively
removes the influence of unlearning samples while maintaining the
representations learned from the remaining samples. Experiments on a variety of
datasets and models on both class unlearning and sample unlearning showed that
contrastive unlearning achieves the best unlearning effects and efficiency with
the lowest performance loss compared with the state-of-the-art algorithms.
Related papers
- Unlearning in- vs. out-of-distribution data in LLMs under gradient-based method [31.268301764230525]
This work formalizes a metric to evaluate unlearning quality in generative models.
We use it to assess the trade-offs between unlearning quality and performance.
We further evaluate how example's memorization and difficulty affect unlearning under a classical gradient ascent-based approach.
arXiv Detail & Related papers (2024-11-07T03:02:09Z) - Model Sparsity Can Simplify Machine Unlearning [33.18951938708467]
In response to recent data regulation requirements, machine unlearning (MU) has emerged as a critical process.
Our study introduces a novel model-based perspective: model sparsification via weight pruning.
We show in both theory and practice that model sparsity can boost the multi-criteria unlearning performance of an approximate unlearner.
arXiv Detail & Related papers (2023-04-11T02:12:02Z) - Mind Your Outliers! Investigating the Negative Impact of Outliers on
Active Learning for Visual Question Answering [71.15403434929915]
We show that across 5 models and 4 datasets on the task of visual question answering, a wide variety of active learning approaches fail to outperform random selection.
We identify the problem as collective outliers -- groups of examples that active learning methods prefer to acquire but models fail to learn.
We show that active learning sample efficiency increases significantly as the number of collective outliers in the active learning pool decreases.
arXiv Detail & Related papers (2021-07-06T00:52:11Z) - Incremental False Negative Detection for Contrastive Learning [95.68120675114878]
We introduce a novel incremental false negative detection for self-supervised contrastive learning.
During contrastive learning, we discuss two strategies to explicitly remove the detected false negatives.
Our proposed method outperforms other self-supervised contrastive learning frameworks on multiple benchmarks within a limited compute.
arXiv Detail & Related papers (2021-06-07T15:29:14Z) - Self-Damaging Contrastive Learning [92.34124578823977]
Unlabeled data in reality is commonly imbalanced and shows a long-tail distribution.
This paper proposes a principled framework called Self-Damaging Contrastive Learning to automatically balance the representation learning without knowing the classes.
Our experiments show that SDCLR significantly improves not only overall accuracies but also balancedness.
arXiv Detail & Related papers (2021-06-06T00:04:49Z) - Solving Inefficiency of Self-supervised Representation Learning [87.30876679780532]
Existing contrastive learning methods suffer from very low learning efficiency.
Under-clustering and over-clustering problems are major obstacles to learning efficiency.
We propose a novel self-supervised learning framework using a median triplet loss.
arXiv Detail & Related papers (2021-04-18T07:47:10Z) - Doubly Contrastive Deep Clustering [135.7001508427597]
We present a novel Doubly Contrastive Deep Clustering (DCDC) framework, which constructs contrastive loss over both sample and class views.
Specifically, for the sample view, we set the class distribution of the original sample and its augmented version as positive sample pairs.
For the class view, we build the positive and negative pairs from the sample distribution of the class.
In this way, two contrastive losses successfully constrain the clustering results of mini-batch samples in both sample and class level.
arXiv Detail & Related papers (2021-03-09T15:15:32Z) - Coded Machine Unlearning [34.08435990347253]
We present a coded learning protocol where the dataset is linearly coded before the learning phase.
We also present the corresponding unlearning protocol for the coded learning model along with a discussion on the proposed protocol's success in ensuring perfect unlearning.
arXiv Detail & Related papers (2020-12-31T17:20:34Z) - Contrastive Learning with Hard Negative Samples [80.12117639845678]
We develop a new family of unsupervised sampling methods for selecting hard negative samples.
A limiting case of this sampling results in a representation that tightly clusters each class, and pushes different classes as far apart as possible.
The proposed method improves downstream performance across multiple modalities, requires only few additional lines of code to implement, and introduces no computational overhead.
arXiv Detail & Related papers (2020-10-09T14:18:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.