GCR: Gradient Coreset Based Replay Buffer Selection For Continual
Learning
- URL: http://arxiv.org/abs/2111.11210v1
- Date: Thu, 18 Nov 2021 18:01:57 GMT
- Title: GCR: Gradient Coreset Based Replay Buffer Selection For Continual
Learning
- Authors: Rishabh Tiwari, Krishnateja Killamsetty, Rishabh Iyer, Pradeep Shenoy
- Abstract summary: We show significant gains (2%-4% absolute) over the state-of-the-art in the well-studied offline continual learning setting.
Our findings also effectively transfer to online / streaming CL settings, showing upto 5% gains over existing approaches.
- Score: 1.911678487931003
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Continual learning (CL) aims to develop techniques by which a single model
adapts to an increasing number of tasks encountered sequentially, thereby
potentially leveraging learnings across tasks in a resource-efficient manner. A
major challenge for CL systems is catastrophic forgetting, where earlier tasks
are forgotten while learning a new task. To address this, replay-based CL
approaches maintain and repeatedly retrain on a small buffer of data selected
across encountered tasks. We propose Gradient Coreset Replay (GCR), a novel
strategy for replay buffer selection and update using a carefully designed
optimization criterion. Specifically, we select and maintain a "coreset" that
closely approximates the gradient of all the data seen so far with respect to
current model parameters, and discuss key strategies needed for its effective
application to the continual learning setting. We show significant gains (2%-4%
absolute) over the state-of-the-art in the well-studied offline continual
learning setting. Our findings also effectively transfer to online / streaming
CL settings, showing upto 5% gains over existing approaches. Finally, we
demonstrate the value of supervised contrastive loss for continual learning,
which yields a cumulative gain of up to 5% accuracy when combined with our
subset selection strategy.
Related papers
- Adaptive Rentention & Correction for Continual Learning [114.5656325514408]
A common problem in continual learning is the classification layer's bias towards the most recent task.
We name our approach Adaptive Retention & Correction (ARC)
ARC achieves an average performance increase of 2.7% and 2.6% on the CIFAR-100 and Imagenet-R datasets.
arXiv Detail & Related papers (2024-05-23T08:43:09Z) - CORE: Mitigating Catastrophic Forgetting in Continual Learning through Cognitive Replay [14.169588352209907]
COgnitive REplay (CORE) draws inspiration from human cognitive review processes.
CORE achieves an average accuracy of 37.95% on split-CIFAR10, surpassing the best baseline method by 6.52%.
It significantly enhances the accuracy of the poorest-performing task by 6.30% compared to the top baseline.
arXiv Detail & Related papers (2024-02-02T12:04:44Z) - RanPAC: Random Projections and Pre-trained Models for Continual Learning [59.07316955610658]
Continual learning (CL) aims to learn different tasks (such as classification) in a non-stationary data stream without forgetting old ones.
We propose a concise and effective approach for CL with pre-trained models.
arXiv Detail & Related papers (2023-07-05T12:49:02Z) - Mitigating Catastrophic Forgetting in Task-Incremental Continual
Learning with Adaptive Classification Criterion [50.03041373044267]
We propose a Supervised Contrastive learning framework with adaptive classification criterion for Continual Learning.
Experiments show that CFL achieves state-of-the-art performance and has a stronger ability to overcome compared with the classification baselines.
arXiv Detail & Related papers (2023-05-20T19:22:40Z) - Gradient-Matching Coresets for Rehearsal-Based Continual Learning [6.243028964381449]
The goal of continual learning (CL) is to efficiently update a machine learning model with new data without forgetting previously-learned knowledge.
Most widely-used CL methods rely on a rehearsal memory of data points to be reused while training on new data.
We devise a coreset selection method for rehearsal-based continual learning.
arXiv Detail & Related papers (2022-03-28T07:37:17Z) - Learning Bayesian Sparse Networks with Full Experience Replay for
Continual Learning [54.7584721943286]
Continual Learning (CL) methods aim to enable machine learning models to learn new tasks without catastrophic forgetting of those that have been previously mastered.
Existing CL approaches often keep a buffer of previously-seen samples, perform knowledge distillation, or use regularization techniques towards this goal.
We propose to only activate and select sparse neurons for learning current and past tasks at any stage.
arXiv Detail & Related papers (2022-02-21T13:25:03Z) - vCLIMB: A Novel Video Class Incremental Learning Benchmark [53.90485760679411]
We introduce vCLIMB, a novel video continual learning benchmark.
vCLIMB is a standardized test-bed to analyze catastrophic forgetting of deep models in video continual learning.
We propose a temporal consistency regularization that can be applied on top of memory-based continual learning methods.
arXiv Detail & Related papers (2022-01-23T22:14:17Z) - Online Coreset Selection for Rehearsal-based Continual Learning [65.85595842458882]
In continual learning, we store a subset of training examples (coreset) to be replayed later to alleviate catastrophic forgetting.
We propose Online Coreset Selection (OCS), a simple yet effective method that selects the most representative and informative coreset at each iteration.
Our proposed method maximizes the model's adaptation to a target dataset while selecting high-affinity samples to past tasks, which directly inhibits catastrophic forgetting.
arXiv Detail & Related papers (2021-06-02T11:39:25Z) - Continual Learning From Unlabeled Data Via Deep Clustering [7.704949298975352]
Continual learning aims to learn new tasks incrementally using less computation and memory resources instead of retraining the model from scratch whenever new task arrives.
We introduce a new framework to make continual learning feasible in unsupervised mode by using pseudo label obtained from cluster assignments to update model.
arXiv Detail & Related papers (2021-04-14T23:46:17Z) - Continual Learning with Node-Importance based Adaptive Group Sparse
Regularization [30.23319528662881]
We propose a novel regularization-based continual learning method, dubbed as Adaptive Group Sparsity based Continual Learning (AGS-CL)
Our method selectively employs the two penalties when learning each node based its the importance, which is adaptively updated after learning each new task.
arXiv Detail & Related papers (2020-03-30T18:21:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.