Learning with Recoverable Forgetting
- URL: http://arxiv.org/abs/2207.08224v1
- Date: Sun, 17 Jul 2022 16:42:31 GMT
- Title: Learning with Recoverable Forgetting
- Authors: Jingwen Ye, Yifang Fu, Jie Song, Xingyi Yang, Songhua Liu, Xin Jin,
Mingli Song, Xinchao Wang
- Abstract summary: Learning wIth Recoverable Forgetting explicitly handles the task- or sample-specific knowledge removal and recovery.
Specifically, LIRF brings in two innovative schemes, namely knowledge deposit and withdrawal.
We conduct experiments on several datasets, and demonstrate that the proposed LIRF strategy yields encouraging results with gratifying generalization capability.
- Score: 77.56338597012927
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Life-long learning aims at learning a sequence of tasks without forgetting
the previously acquired knowledge. However, the involved training data may not
be life-long legitimate due to privacy or copyright reasons. In practical
scenarios, for instance, the model owner may wish to enable or disable the
knowledge of specific tasks or specific samples from time to time. Such
flexible control over knowledge transfer, unfortunately, has been largely
overlooked in previous incremental or decremental learning methods, even at a
problem-setup level. In this paper, we explore a novel learning scheme, termed
as Learning wIth Recoverable Forgetting (LIRF), that explicitly handles the
task- or sample-specific knowledge removal and recovery. Specifically, LIRF
brings in two innovative schemes, namely knowledge deposit and withdrawal,
which allow for isolating user-designated knowledge from a pre-trained network
and injecting it back when necessary. During the knowledge deposit process, the
specified knowledge is extracted from the target network and stored in a
deposit module, while the insensitive or general knowledge of the target
network is preserved and further augmented. During knowledge withdrawal, the
taken-off knowledge is added back to the target network. The deposit and
withdraw processes only demand for a few epochs of finetuning on the removal
data, ensuring both data and time efficiency. We conduct experiments on several
datasets, and demonstrate that the proposed LIRF strategy yields encouraging
results with gratifying generalization capability.
Related papers
- Extracting Unlearned Information from LLMs with Activation Steering [46.16882599881247]
Unlearning has emerged as a solution to remove sensitive knowledge from models after training.
We propose activation steering as a method for exact information retrieval from unlearned models.
Our results demonstrate that exact information retrieval from unlearned models is possible, highlighting a severe vulnerability of current unlearning techniques.
arXiv Detail & Related papers (2024-11-04T21:42:56Z) - Hierarchical Prompts for Rehearsal-free Continual Learning [67.37739666753008]
Continual learning endeavors to equip the model with the capability to integrate current task knowledge while mitigating the forgetting of past task knowledge.
Inspired by prompt tuning, prompt-based methods maintain a frozen backbone and train with slight learnable prompts.
This paper introduces a novel rehearsal-free paradigm for continual learning termed Hierarchical Prompts (H-Prompts)
arXiv Detail & Related papers (2024-01-21T16:59:44Z) - Learning to Retain while Acquiring: Combating Distribution-Shift in
Adversarial Data-Free Knowledge Distillation [31.294947552032088]
Data-free Knowledge Distillation (DFKD) has gained popularity recently, with the fundamental idea of carrying out knowledge transfer from a Teacher to a Student neural network in the absence of training data.
We propose a meta-learning inspired framework by treating the task of Knowledge-Acquisition (learning from newly generated samples) and Knowledge-Retention (retaining knowledge on previously met samples) as meta-train and meta-test.
arXiv Detail & Related papers (2023-02-28T03:50:56Z) - Beyond Not-Forgetting: Continual Learning with Backward Knowledge
Transfer [39.99577526417276]
In continual learning (CL) an agent can improve the learning performance of both a new task and old' tasks.
Most existing CL methods focus on addressing catastrophic forgetting in neural networks by minimizing the modification of the learnt model for old tasks.
We propose a new CL method with Backward knowlEdge tRansfer (CUBER) for a fixed capacity neural network without data replay.
arXiv Detail & Related papers (2022-11-01T23:55:51Z) - A Unified End-to-End Retriever-Reader Framework for Knowledge-based VQA [67.75989848202343]
This paper presents a unified end-to-end retriever-reader framework towards knowledge-based VQA.
We shed light on the multi-modal implicit knowledge from vision-language pre-training models to mine its potential in knowledge reasoning.
Our scheme is able to not only provide guidance for knowledge retrieval, but also drop these instances potentially error-prone towards question answering.
arXiv Detail & Related papers (2022-06-30T02:35:04Z) - Transferability in Deep Learning: A Survey [80.67296873915176]
The ability to acquire and reuse knowledge is known as transferability in deep learning.
We present this survey to connect different isolated areas in deep learning with their relation to transferability.
We implement a benchmark and an open-source library, enabling a fair evaluation of deep learning methods in terms of transferability.
arXiv Detail & Related papers (2022-01-15T15:03:17Z) - Preserving Earlier Knowledge in Continual Learning with the Help of All
Previous Feature Extractors [63.21036904487014]
Continual learning of new knowledge over time is one desirable capability for intelligent systems to recognize more and more classes of objects.
We propose a simple yet effective fusion mechanism by including all the previously learned feature extractors into the intelligent model.
Experiments on multiple classification tasks show that the proposed approach can effectively reduce the forgetting of old knowledge, achieving state-of-the-art continual learning performance.
arXiv Detail & Related papers (2021-04-28T07:49:24Z) - Data-free Knowledge Distillation for Segmentation using Data-Enriching
GAN [0.0]
We propose a new training framework for performing knowledge distillation in a data-free setting.
We get an improvement of 6.93% in Mean IoU over previous approaches.
arXiv Detail & Related papers (2020-11-02T08:16:42Z) - Bilevel Continual Learning [76.50127663309604]
We present a novel framework of continual learning named "Bilevel Continual Learning" (BCL)
Our experiments on continual learning benchmarks demonstrate the efficacy of the proposed BCL compared to many state-of-the-art methods.
arXiv Detail & Related papers (2020-07-30T16:00:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.