Balancing the Causal Effects in Class-Incremental Learning
- URL: http://arxiv.org/abs/2402.10063v1
- Date: Thu, 15 Feb 2024 16:30:45 GMT
- Title: Balancing the Causal Effects in Class-Incremental Learning
- Authors: Junhao Zheng, Ruiyan Wang, Chongzhi Zhang, Huawen Feng, Qianli Ma
- Abstract summary: Class-Incremental Learning (CIL) is a practical and challenging problem for achieving general artificial intelligence.
We show that the crux lies in the imbalanced causal effects between new and old data.
We propose Balancing the Causal Effects (BaCE) in CIL to alleviate this problem.
- Score: 23.35478989162079
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Class-Incremental Learning (CIL) is a practical and challenging problem for
achieving general artificial intelligence. Recently, Pre-Trained Models (PTMs)
have led to breakthroughs in both visual and natural language processing tasks.
Despite recent studies showing PTMs' potential ability to learn sequentially, a
plethora of work indicates the necessity of alleviating the catastrophic
forgetting of PTMs. Through a pilot study and a causal analysis of CIL, we
reveal that the crux lies in the imbalanced causal effects between new and old
data. Specifically, the new data encourage models to adapt to new classes while
hindering the adaptation of old classes. Similarly, the old data encourages
models to adapt to old classes while hindering the adaptation of new classes.
In other words, the adaptation process between new and old classes conflicts
from the causal perspective. To alleviate this problem, we propose Balancing
the Causal Effects (BaCE) in CIL. Concretely, BaCE proposes two objectives for
building causal paths from both new and old data to the prediction of new and
classes, respectively. In this way, the model is encouraged to adapt to all
classes with causal effects from both new and old data and thus alleviates the
causal imbalance problem. We conduct extensive experiments on continual image
classification, continual text classification, and continual named entity
recognition. Empirical results show that BaCE outperforms a series of CIL
methods on different tasks and settings.
Related papers
- CSTA: Spatial-Temporal Causal Adaptive Learning for Exemplar-Free Video Class-Incremental Learning [62.69917996026769]
A class-incremental learning task requires learning and preserving both spatial appearance and temporal action involvement.
We propose a framework that equips separate adapters to learn new class patterns, accommodating the incremental information requirements unique to each class.
A causal compensation mechanism is proposed to reduce the conflicts during increment and memorization for between different types of information.
arXiv Detail & Related papers (2025-01-13T11:34:55Z) - Happy: A Debiased Learning Framework for Continual Generalized Category Discovery [54.54153155039062]
This paper explores the underexplored task of Continual Generalized Category Discovery (C-GCD)
C-GCD aims to incrementally discover new classes from unlabeled data while maintaining the ability to recognize previously learned classes.
We introduce a debiased learning framework, namely Happy, characterized by Hardness-aware prototype sampling and soft entropy regularization.
arXiv Detail & Related papers (2024-10-09T04:18:51Z) - Few-Shot Class-Incremental Learning with Prior Knowledge [94.95569068211195]
We propose Learning with Prior Knowledge (LwPK) to enhance the generalization ability of the pre-trained model.
Experimental results indicate that LwPK effectively enhances the model resilience against catastrophic forgetting.
arXiv Detail & Related papers (2024-02-02T08:05:35Z) - Continual Learning with Pre-Trained Models: A Survey [61.97613090666247]
Continual Learning aims to overcome the catastrophic forgetting of former knowledge when learning new ones.
This paper presents a comprehensive survey of the latest advancements in PTM-based CL.
arXiv Detail & Related papers (2024-01-29T18:27:52Z) - Class-Incremental Learning of Plant and Disease Detection: Growing
Branches with Knowledge Distillation [0.0]
This paper investigates the problem of class-incremental object detection for agricultural applications.
We adapt two public datasets to include new categories over time, simulating a more realistic and dynamic scenario.
We compare three class-incremental learning methods that leverage different forms of knowledge distillation to mitigate catastrophic forgetting.
arXiv Detail & Related papers (2023-04-13T15:40:41Z) - Class-Incremental Learning: A Survey [84.30083092434938]
Class-Incremental Learning (CIL) enables the learner to incorporate the knowledge of new classes incrementally.
CIL tends to catastrophically forget the characteristics of former ones, and its performance drastically degrades.
We provide a rigorous and unified evaluation of 17 methods in benchmark image classification tasks to find out the characteristics of different algorithms.
arXiv Detail & Related papers (2023-02-07T17:59:05Z) - Multi-Granularity Regularized Re-Balancing for Class Incremental
Learning [32.52884416761171]
Deep learning models suffer from catastrophic forgetting when learning new tasks.
Data imbalance between old and new classes is a key issue that leads to performance degradation of the model.
We propose an assumption-agnostic method, Multi-Granularity Regularized re-Balancing, to address this problem.
arXiv Detail & Related papers (2022-06-30T11:04:51Z) - Multi-View Correlation Distillation for Incremental Object Detection [12.536640582318949]
We propose a novel textbfMulti-textbfView textbfCorrelation textbfDistillation (MVCD) based incremental object detection method.
arXiv Detail & Related papers (2021-07-05T04:36:33Z) - Class-incremental Learning with Rectified Feature-Graph Preservation [24.098892115785066]
A central theme of this paper is to learn new classes that arrive in sequential phases over time.
We propose a weighted-Euclidean regularization for old knowledge preservation.
We show how it can work with binary cross-entropy to increase class separation for effective learning of new classes.
arXiv Detail & Related papers (2020-12-15T07:26:04Z) - Learning Adaptive Embedding Considering Incremental Class [55.21855842960139]
Class-Incremental Learning (CIL) aims to train a reliable model with the streaming data, which emerges unknown classes sequentially.
Different from traditional closed set learning, CIL has two main challenges: 1) Novel class detection.
After the novel classes are detected, the model needs to be updated without re-training using entire previous data.
arXiv Detail & Related papers (2020-08-31T04:11:24Z) - Self-Supervised Learning Aided Class-Incremental Lifelong Learning [17.151579393716958]
We study the issue of catastrophic forgetting in class-incremental learning (Class-IL)
In training procedure of Class-IL, as the model has no knowledge about following tasks, it would only extract features necessary for tasks learned so far, whose information is insufficient for joint classification.
We propose to combine self-supervised learning, which can provide effective representations without requiring labels, with Class-IL to partly get around this problem.
arXiv Detail & Related papers (2020-06-10T15:15:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.