Consistency is the key to further mitigating catastrophic forgetting in
continual learning
- URL: http://arxiv.org/abs/2207.04998v1
- Date: Mon, 11 Jul 2022 16:44:49 GMT
- Title: Consistency is the key to further mitigating catastrophic forgetting in
continual learning
- Authors: Prashant Bhat, Bahram Zonooz, Elahe Arani
- Abstract summary: Experience Replay (ER) does not perform well under low-buffer regimes and longer task sequences.
consistency in predictions of soft-targets can assist ER in preserving information pertaining to previous tasks better.
We propose to cast consistency regularization as a self-supervised pretext task.
- Score: 14.674494335647841
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Deep neural networks struggle to continually learn multiple sequential tasks
due to catastrophic forgetting of previously learned tasks. Rehearsal-based
methods which explicitly store previous task samples in the buffer and
interleave them with the current task samples have proven to be the most
effective in mitigating forgetting. However, Experience Replay (ER) does not
perform well under low-buffer regimes and longer task sequences as its
performance is commensurate with the buffer size. Consistency in predictions of
soft-targets can assist ER in preserving information pertaining to previous
tasks better as soft-targets capture the rich similarity structure of the data.
Therefore, we examine the role of consistency regularization in ER framework
under various continual learning scenarios. We also propose to cast consistency
regularization as a self-supervised pretext task thereby enabling the use of a
wide variety of self-supervised learning methods as regularizers. While
simultaneously enhancing model calibration and robustness to natural
corruptions, regularizing consistency in predictions results in lesser
forgetting across all continual learning scenarios. Among the different
families of regularizers, we find that stricter consistency constraints
preserve previous task information in ER better.
Related papers
- Investigating the Interplay of Prioritized Replay and Generalization [23.248982121562985]
We study Prioritized Experience Replay (PER), where sampling is done proportionally to TD errors.
PER is inspired by the success of prioritized sweeping in dynamic programming.
arXiv Detail & Related papers (2024-07-12T21:56:24Z) - Reshaping the Online Data Buffering and Organizing Mechanism for Continual Test-Time Adaptation [49.53202761595912]
Continual Test-Time Adaptation involves adapting a pre-trained source model to continually changing unsupervised target domains.
We analyze the challenges of this task: online environment, unsupervised nature, and the risks of error accumulation and catastrophic forgetting.
We propose an uncertainty-aware buffering approach to identify and aggregate significant samples with high certainty from the unsupervised, single-pass data stream.
arXiv Detail & Related papers (2024-07-12T15:48:40Z) - Enhancing Robustness of Vision-Language Models through Orthogonality Learning and Self-Regularization [77.62516752323207]
We introduce an orthogonal fine-tuning method for efficiently fine-tuning pretrained weights and enabling enhanced robustness and generalization.
A self-regularization strategy is further exploited to maintain the stability in terms of zero-shot generalization of VLMs, dubbed OrthSR.
For the first time, we revisit the CLIP and CoOp with our method to effectively improve the model on few-shot image classficiation scenario.
arXiv Detail & Related papers (2024-07-11T10:35:53Z) - Selective Learning: Towards Robust Calibration with Dynamic Regularization [79.92633587914659]
Miscalibration in deep learning refers to there is a discrepancy between the predicted confidence and performance.
We introduce Dynamic Regularization (DReg) which aims to learn what should be learned during training thereby circumventing the confidence adjusting trade-off.
arXiv Detail & Related papers (2024-02-13T11:25:20Z) - Enhancing Consistency and Mitigating Bias: A Data Replay Approach for
Incremental Learning [100.7407460674153]
Deep learning systems are prone to catastrophic forgetting when learning from a sequence of tasks.
To mitigate the problem, a line of methods propose to replay the data of experienced tasks when learning new tasks.
However, it is not expected in practice considering the memory constraint or data privacy issue.
As a replacement, data-free data replay methods are proposed by inverting samples from the classification model.
arXiv Detail & Related papers (2024-01-12T12:51:12Z) - Noisy Correspondence Learning with Self-Reinforcing Errors Mitigation [63.180725016463974]
Cross-modal retrieval relies on well-matched large-scale datasets that are laborious in practice.
We introduce a novel noisy correspondence learning framework, namely textbfSelf-textbfReinforcing textbfErrors textbfMitigation (SREM)
arXiv Detail & Related papers (2023-12-27T09:03:43Z) - Replay-enhanced Continual Reinforcement Learning [37.34722105058351]
We introduce RECALL, a replay-enhanced method that greatly improves the plasticity of existing replay-based methods on new tasks.
Experiments on the Continual World benchmark show that RECALL performs significantly better than purely perfect memory replay.
arXiv Detail & Related papers (2023-11-20T06:21:52Z) - Minimax Forward and Backward Learning of Evolving Tasks with Performance
Guarantees [6.008132390640294]
The incremental learning of a growing sequence of tasks holds promise to enable accurate classification.
This paper presents incremental minimax risk classifiers (IMRCs) that effectively exploit forward and backward learning.
IMRCs can result in a significant performance improvement, especially for reduced sample sizes.
arXiv Detail & Related papers (2023-10-24T16:21:41Z) - Towards Robust Continual Learning with Bayesian Adaptive Moment Regularization [51.34904967046097]
Continual learning seeks to overcome the challenge of catastrophic forgetting, where a model forgets previously learnt information.
We introduce a novel prior-based method that better constrains parameter growth, reducing catastrophic forgetting.
Results show that BAdam achieves state-of-the-art performance for prior-based methods on challenging single-headed class-incremental experiments.
arXiv Detail & Related papers (2023-09-15T17:10:51Z) - Self-paced Weight Consolidation for Continual Learning [39.27729549041708]
Continual learning algorithms are popular in preventing catastrophic forgetting in sequential task learning settings.
We propose a self-paced Weight Consolidation (spWC) framework to attain continual learning.
arXiv Detail & Related papers (2023-07-20T13:07:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.