Online Continual Learning under Extreme Memory Constraints
- URL: http://arxiv.org/abs/2008.01510v3
- Date: Wed, 12 Jan 2022 14:09:48 GMT
- Title: Online Continual Learning under Extreme Memory Constraints
- Authors: Enrico Fini, St\'ephane Lathuili\`ere, Enver Sangineto, Moin Nabi,
Elisa Ricci
- Abstract summary: We introduce the novel problem of Memory-Constrained Online Continual Learning (MC-OCL)
MC-OCL imposes strict constraints on the memory overhead that a possible algorithm can use to avoid catastrophic forgetting.
We propose an algorithmic solution to MC-OCL: Batch-level Distillation (BLD), a regularization-based CL approach.
- Score: 40.80045285324969
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Continual Learning (CL) aims to develop agents emulating the human ability to
sequentially learn new tasks while being able to retain knowledge obtained from
past experiences. In this paper, we introduce the novel problem of
Memory-Constrained Online Continual Learning (MC-OCL) which imposes strict
constraints on the memory overhead that a possible algorithm can use to avoid
catastrophic forgetting. As most, if not all, previous CL methods violate these
constraints, we propose an algorithmic solution to MC-OCL: Batch-level
Distillation (BLD), a regularization-based CL approach, which effectively
balances stability and plasticity in order to learn from data streams, while
preserving the ability to solve old tasks through distillation. Our extensive
experimental evaluation, conducted on three publicly available benchmarks,
empirically demonstrates that our approach successfully addresses the MC-OCL
problem and achieves comparable accuracy to prior distillation methods
requiring higher memory overhead.
Related papers
- ICL-TSVD: Bridging Theory and Practice in Continual Learning with Pre-trained Models [103.45785408116146]
Continual learning (CL) aims to train a model that can solve multiple tasks presented sequentially.
Recent CL approaches have achieved strong performance by leveraging large pre-trained models that generalize well to downstream tasks.
However, such methods lack theoretical guarantees, making them prone to unexpected failures.
We bridge this gap by integrating an empirically strong approach into a principled framework, designed to prevent forgetting.
arXiv Detail & Related papers (2024-10-01T12:58:37Z) - Forward-Backward Knowledge Distillation for Continual Clustering [14.234785944941672]
Unsupervised Continual Learning (UCL) is a burgeoning field in machine learning, focusing on enabling neural networks to sequentially learn tasks without explicit label information.
Catastrophic Forgetting (CF) poses a significant challenge in continual learning, especially in UCL, where labeled information of data is not accessible.
We introduce the concept of Unsupervised Continual Clustering (UCC), demonstrating enhanced performance and memory efficiency in clustering across various tasks.
arXiv Detail & Related papers (2024-05-29T16:13:54Z) - Overcoming Domain Drift in Online Continual Learning [24.86094018430407]
Online Continual Learning (OCL) empowers machine learning models to acquire new knowledge online across a sequence of tasks.
OCL faces a significant challenge: catastrophic forgetting, wherein the model learned in previous tasks is substantially overwritten upon encountering new tasks.
We propose a novel rehearsal strategy, Drift-Reducing Rehearsal (DRR), to anchor the domain of old tasks and reduce the negative transfer effects.
arXiv Detail & Related papers (2024-05-15T06:57:18Z) - Continual Learning on a Diet: Learning from Sparsely Labeled Streams Under Constrained Computation [123.4883806344334]
We study a realistic Continual Learning setting where learning algorithms are granted a restricted computational budget per time step while training.
We apply this setting to large-scale semi-supervised Continual Learning scenarios with sparse label rates.
Our extensive analysis and ablations demonstrate that DietCL is stable under a full spectrum of label sparsity, computational budget, and various other ablations.
arXiv Detail & Related papers (2024-04-19T10:10:39Z) - Towards Robust Continual Learning with Bayesian Adaptive Moment Regularization [51.34904967046097]
Continual learning seeks to overcome the challenge of catastrophic forgetting, where a model forgets previously learnt information.
We introduce a novel prior-based method that better constrains parameter growth, reducing catastrophic forgetting.
Results show that BAdam achieves state-of-the-art performance for prior-based methods on challenging single-headed class-incremental experiments.
arXiv Detail & Related papers (2023-09-15T17:10:51Z) - Continual Learning with Dirichlet Generative-based Rehearsal [22.314195832409755]
We present Dirichlet Continual Learning, a novel generative-based rehearsal strategy for task-oriented dialogue systems.
We also introduce Jensen-Shannon Knowledge Distillation (JSKD), a robust logit-based knowledge distillation method.
Our experiments confirm the efficacy of our approach in both intent detection and slot-filling tasks, outperforming state-of-the-art methods.
arXiv Detail & Related papers (2023-09-13T12:30:03Z) - CCL: Continual Contrastive Learning for LiDAR Place Recognition [5.025654873456756]
Current deep learning-based methods suffer from poor generalization ability and catastrophic forgetting.
We propose a continual contrastive learning method, named CCL, to tackle the catastrophic forgetting problem.
Our method consistently improves the performance of different methods in different environments outperforming the state-of-the-art continual learning method.
arXiv Detail & Related papers (2023-03-24T12:14:54Z) - A Study of Continual Learning Methods for Q-Learning [78.6363825307044]
We present an empirical study on the use of continual learning (CL) methods in a reinforcement learning (RL) scenario.
Our results show that dedicated CL methods can significantly improve learning when compared to the baseline technique of "experience replay"
arXiv Detail & Related papers (2022-06-08T14:51:52Z) - Bilevel Continual Learning [76.50127663309604]
We present a novel framework of continual learning named "Bilevel Continual Learning" (BCL)
Our experiments on continual learning benchmarks demonstrate the efficacy of the proposed BCL compared to many state-of-the-art methods.
arXiv Detail & Related papers (2020-07-30T16:00:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.