New Insights on Relieving Task-Recency Bias for Online Class Incremental
Learning
- URL: http://arxiv.org/abs/2302.08243v2
- Date: Wed, 18 Oct 2023 07:50:48 GMT
- Title: New Insights on Relieving Task-Recency Bias for Online Class Incremental
Learning
- Authors: Guoqiang Liang, Zhaojie Chen, Zhaoqiang Chen, Shiyu Ji, Yanning Zhang
- Abstract summary: In all settings, the online class incremental learning (OCIL) is more challenging and can be encountered more frequently in real world.
To strike a preferable trade-off between stability and plasticity, we propose an Adaptive Focus Shifting algorithm.
- Score: 37.888061221999294
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: To imitate the ability of keeping learning of human, continual learning which
can learn from a never-ending data stream has attracted more interests
recently. In all settings, the online class incremental learning (OCIL), where
incoming samples from data stream can be used only once, is more challenging
and can be encountered more frequently in real world. Actually, all continual
learning models face a stability-plasticity dilemma, where the stability means
the ability to preserve old knowledge while the plasticity denotes the ability
to incorporate new knowledge. Although replay-based methods have shown
exceptional promise, most of them concentrate on the strategy for updating and
retrieving memory to keep stability at the expense of plasticity. To strike a
preferable trade-off between stability and plasticity, we propose an Adaptive
Focus Shifting algorithm (AFS), which dynamically adjusts focus to ambiguous
samples and non-target logits in model learning. Through a deep analysis of the
task-recency bias caused by class imbalance, we propose a revised focal loss to
mainly keep stability. \Rt{By utilizing a new weight function, the revised
focal loss will pay more attention to current ambiguous samples, which are the
potentially valuable samples to make model progress quickly.} To promote
plasticity, we introduce a virtual knowledge distillation. By designing a
virtual teacher, it assigns more attention to non-target classes, which can
surmount overconfidence and encourage model to focus on inter-class
information. Extensive experiments on three popular datasets for OCIL have
shown the effectiveness of AFS. The code will be available at
\url{https://github.com/czjghost/AFS}.
Related papers
- Temporal-Difference Variational Continual Learning [89.32940051152782]
A crucial capability of Machine Learning models in real-world applications is the ability to continuously learn new tasks.
In Continual Learning settings, models often struggle to balance learning new tasks with retaining previous knowledge.
We propose new learning objectives that integrate the regularization effects of multiple previous posterior estimations.
arXiv Detail & Related papers (2024-10-10T10:58:41Z) - Learn from the Learnt: Source-Free Active Domain Adaptation via Contrastive Sampling and Visual Persistence [60.37934652213881]
Domain Adaptation (DA) facilitates knowledge transfer from a source domain to a related target domain.
This paper investigates a practical DA paradigm, namely Source data-Free Active Domain Adaptation (SFADA), where source data becomes inaccessible during adaptation.
We present learn from the learnt (LFTL), a novel paradigm for SFADA to leverage the learnt knowledge from the source pretrained model and actively iterated models without extra overhead.
arXiv Detail & Related papers (2024-07-26T17:51:58Z) - Branch-Tuning: Balancing Stability and Plasticity for Continual Self-Supervised Learning [33.560003528712414]
Self-supervised learning (SSL) has emerged as an effective paradigm for deriving general representations from vast amounts of unlabeled data.
This poses a challenge in striking a balance between stability and plasticity when adapting to new information.
We propose Branch-tuning, an efficient and straightforward method that achieves a balance between stability and plasticity in continual SSL.
arXiv Detail & Related papers (2024-03-27T05:38:48Z) - Learning Objective-Specific Active Learning Strategies with Attentive
Neural Processes [72.75421975804132]
Learning Active Learning (LAL) suggests to learn the active learning strategy itself, allowing it to adapt to the given setting.
We propose a novel LAL method for classification that exploits symmetry and independence properties of the active learning problem.
Our approach is based on learning from a myopic oracle, which gives our model the ability to adapt to non-standard objectives.
arXiv Detail & Related papers (2023-09-11T14:16:37Z) - SRIL: Selective Regularization for Class-Incremental Learning [5.810252620242912]
Class-Incremental Learning aims to create an integrated model that balances plasticity and stability to overcome this challenge.
We propose a selective regularization method that accepts new knowledge while maintaining previous knowledge.
We validate the effectiveness of the proposed method through extensive experimental protocols using CIFAR-100, ImageNet-Subset, and ImageNet-Full.
arXiv Detail & Related papers (2023-05-09T05:04:35Z) - New Insights for the Stability-Plasticity Dilemma in Online Continual
Learning [21.664470275289407]
We propose an online continual learning framework named multi-scale feature adaptation network (MuFAN)
MuFAN outperforms other state-of-the-art continual learning methods on the SVHN, CIFAR100, miniImageNet, and CORe50 datasets.
arXiv Detail & Related papers (2023-02-17T07:43:59Z) - Adaptively Integrated Knowledge Distillation and Prediction Uncertainty
for Continual Learning [71.43841235954453]
Current deep learning models often suffer from catastrophic forgetting of old knowledge when continually learning new knowledge.
Existing strategies to alleviate this issue often fix the trade-off between keeping old knowledge (stability) and learning new knowledge (plasticity)
arXiv Detail & Related papers (2023-01-18T05:36:06Z) - Mitigating Forgetting in Online Continual Learning via Contrasting
Semantically Distinct Augmentations [22.289830907729705]
Online continual learning (OCL) aims to enable model learning from a non-stationary data stream to continuously acquire new knowledge as well as retain the learnt one.
Main challenge comes from the "catastrophic forgetting" issue -- the inability to well remember the learnt knowledge while learning the new ones.
arXiv Detail & Related papers (2022-11-10T05:29:43Z) - Towards Better Plasticity-Stability Trade-off in Incremental Learning: A
simple Linear Connector [8.13916229438606]
Plasticity-stability dilemma is a main problem for incremental learning.
We show that a simple averaging of two independently optimized optima of networks, null-space projection for past tasks and simple SGD for the current task, can attain a meaningful balance between preserving already learned knowledge and granting sufficient flexibility for learning a new task.
arXiv Detail & Related papers (2021-10-15T07:37:20Z) - Online Continual Learning with Natural Distribution Shifts: An Empirical
Study with Visual Data [101.6195176510611]
"Online" continual learning enables evaluating both information retention and online learning efficacy.
In online continual learning, each incoming small batch of data is first used for testing and then added to the training set, making the problem truly online.
We introduce a new benchmark for online continual visual learning that exhibits large scale and natural distribution shifts.
arXiv Detail & Related papers (2021-08-20T06:17:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.