Generative Feature Replay with Orthogonal Weight Modification for
Continual Learning
- URL: http://arxiv.org/abs/2005.03490v3
- Date: Sat, 12 Sep 2020 03:35:24 GMT
- Title: Generative Feature Replay with Orthogonal Weight Modification for
Continual Learning
- Authors: Gehui Shen, Song Zhang, Xiang Chen and Zhi-Hong Deng
- Abstract summary: generative replay is a promising strategy which generates and replays pseudo data for previous tasks to alleviate catastrophic forgetting.
We propose to replay penultimate layer feature with a generative model; 2) leverage a self-supervised auxiliary task to further enhance the stability of feature.
Empirical results on several datasets show our method always achieves substantial improvement over powerful OWM.
- Score: 20.8966035274874
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The ability of intelligent agents to learn and remember multiple tasks
sequentially is crucial to achieving artificial general intelligence. Many
continual learning (CL) methods have been proposed to overcome catastrophic
forgetting which results from non i.i.d data in the sequential learning of
neural networks. In this paper we focus on class incremental learning, a
challenging CL scenario. For this scenario, generative replay is a promising
strategy which generates and replays pseudo data for previous tasks to
alleviate catastrophic forgetting. However, it is hard to train a generative
model continually for relatively complex data. Based on recently proposed
orthogonal weight modification (OWM) algorithm which can approximately keep
previously learned feature invariant when learning new tasks, we propose to 1)
replay penultimate layer feature with a generative model; 2) leverage a
self-supervised auxiliary task to further enhance the stability of feature.
Empirical results on several datasets show our method always achieves
substantial improvement over powerful OWM while conventional generative replay
always results in a negative effect. Meanwhile our method beats several strong
baselines including one based on real data storage. In addition, we conduct
experiments to study why our method is effective.
Related papers
- Continual Diffuser (CoD): Mastering Continual Offline Reinforcement Learning with Experience Rehearsal [54.93261535899478]
In real-world applications, such as robotic control of reinforcement learning, the tasks are changing, and new tasks arise in a sequential order.
This situation poses the new challenge of plasticity-stability trade-off for training an agent who can adapt to task changes and retain acquired knowledge.
We propose a rehearsal-based continual diffusion model, called Continual diffuser (CoD), to endow the diffuser with the capabilities of quick adaptation (plasticity) and lasting retention (stability)
arXiv Detail & Related papers (2024-09-04T08:21:47Z) - Adaptive Rentention & Correction for Continual Learning [114.5656325514408]
A common problem in continual learning is the classification layer's bias towards the most recent task.
We name our approach Adaptive Retention & Correction (ARC)
ARC achieves an average performance increase of 2.7% and 2.6% on the CIFAR-100 and Imagenet-R datasets.
arXiv Detail & Related papers (2024-05-23T08:43:09Z) - Enhancing Consistency and Mitigating Bias: A Data Replay Approach for
Incremental Learning [100.7407460674153]
Deep learning systems are prone to catastrophic forgetting when learning from a sequence of tasks.
To mitigate the problem, a line of methods propose to replay the data of experienced tasks when learning new tasks.
However, it is not expected in practice considering the memory constraint or data privacy issue.
As a replacement, data-free data replay methods are proposed by inverting samples from the classification model.
arXiv Detail & Related papers (2024-01-12T12:51:12Z) - Prior-Free Continual Learning with Unlabeled Data in the Wild [24.14279172551939]
We propose a Prior-Free Continual Learning (PFCL) method to incrementally update a trained model on new tasks.
PFCL learns new tasks without knowing the task identity or any previous data.
Our experiments show that our PFCL method significantly mitigates forgetting in all three learning scenarios.
arXiv Detail & Related papers (2023-10-16T13:59:56Z) - Complementary Learning Subnetworks for Parameter-Efficient
Class-Incremental Learning [40.13416912075668]
We propose a rehearsal-free CIL approach that learns continually via the synergy between two Complementary Learning Subnetworks.
Our method achieves competitive results against state-of-the-art methods, especially in accuracy gain, memory cost, training efficiency, and task-order.
arXiv Detail & Related papers (2023-06-21T01:43:25Z) - Accelerating Batch Active Learning Using Continual Learning Techniques [5.514154947870141]
A major problem with Active Learning (AL) is high training costs since models are typically retrained from scratch after every query round.
We develop a new class of techniques, circumventing this problem, by biasing further training towards previously labeled sets.
We conduct experiments across many data domains, including natural language, vision, medical imaging, and computational biology.
arXiv Detail & Related papers (2023-05-10T18:34:05Z) - Prompt Conditioned VAE: Enhancing Generative Replay for Lifelong
Learning in Task-Oriented Dialogue [80.05509768165135]
generative replay methods are widely employed to consolidate past knowledge with generated pseudo samples.
Most existing generative replay methods use only a single task-specific token to control their models.
We propose a novel method, prompt conditioned VAE for lifelong learning, to enhance generative replay by incorporating tasks' statistics.
arXiv Detail & Related papers (2022-10-14T13:12:14Z) - Learning Bayesian Sparse Networks with Full Experience Replay for
Continual Learning [54.7584721943286]
Continual Learning (CL) methods aim to enable machine learning models to learn new tasks without catastrophic forgetting of those that have been previously mastered.
Existing CL approaches often keep a buffer of previously-seen samples, perform knowledge distillation, or use regularization techniques towards this goal.
We propose to only activate and select sparse neurons for learning current and past tasks at any stage.
arXiv Detail & Related papers (2022-02-21T13:25:03Z) - Generative Feature Replay For Class-Incremental Learning [46.88667212214957]
We consider a class-incremental setting which means that the task-ID is unknown at inference time.
The imbalance between old and new classes typically results in a bias of the network towards the newest ones.
We propose a solution based on generative feature replay which does not require any exemplars.
arXiv Detail & Related papers (2020-04-20T10:58:20Z) - Generalized Hindsight for Reinforcement Learning [154.0545226284078]
We argue that low-reward data collected while trying to solve one task provides little to no signal for solving that particular task.
We present Generalized Hindsight: an approximate inverse reinforcement learning technique for relabeling behaviors with the right tasks.
arXiv Detail & Related papers (2020-02-26T18:57:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.