TaE: Task-aware Expandable Representation for Long Tail Class
Incremental Learning
- URL: http://arxiv.org/abs/2402.05797v1
- Date: Thu, 8 Feb 2024 16:37:04 GMT
- Title: TaE: Task-aware Expandable Representation for Long Tail Class
Incremental Learning
- Authors: Linjie Li, S. Liu, Zhenyu Wu, JI yang
- Abstract summary: Class-incremental learning (CIL) aims to train classifiers that learn new classes without forgetting old ones.
We introduce a novel Task-aware Expandable (TaE) framework, dynamically allocating and updating task-specific trainable parameters.
We develop a Centroid-Enhanced (CEd) method to guide the update of these task-aware parameters.
- Score: 40.099293186301225
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Class-incremental learning (CIL) aims to train classifiers that learn new
classes without forgetting old ones. Most CIL methods focus on balanced data
distribution for each task, overlooking real-world long-tailed distributions.
Therefore, Long-Tailed Class-Incremental Learning (LT-CIL) has been introduced,
which trains on data where head classes have more samples than tail classes.
Existing methods mainly focus on preserving representative samples from
previous classes to combat catastrophic forgetting. Recently, dynamic network
algorithms frozen old network structures and expanded new ones, achieving
significant performance. However, with the introduction of the long-tail
problem, merely extending task-specific parameters can lead to miscalibrated
predictions, while expanding the entire model results in an explosion of memory
size. To address these issues, we introduce a novel Task-aware Expandable (TaE)
framework, dynamically allocating and updating task-specific trainable
parameters to learn diverse representations from each incremental task, while
resisting forgetting through the majority of frozen model parameters. To
further encourage the class-specific feature representation, we develop a
Centroid-Enhanced (CEd) method to guide the update of these task-aware
parameters. This approach is designed to adaptively minimize the distances
between intra-class features while simultaneously maximizing the distances
between inter-class features across all seen classes. The utility of this
centroid-enhanced method extends to all "training from scratch" CIL algorithms.
Extensive experiments were conducted on CIFAR-100 and ImageNet100 under
different settings, which demonstrates that TaE achieves state-of-the-art
performance.
Related papers
- Efficient Non-Exemplar Class-Incremental Learning with Retrospective Feature Synthesis [21.348252135252412]
Current Non-Exemplar Class-Incremental Learning (NECIL) methods mitigate forgetting by storing a single prototype per class.
We propose a more efficient NECIL method that replaces prototypes with synthesized retrospective features for old classes.
Our method significantly improves the efficiency of non-exemplar class-incremental learning and achieves state-of-the-art performance.
arXiv Detail & Related papers (2024-11-03T07:19:11Z) - Robust Feature Learning and Global Variance-Driven Classifier Alignment
for Long-Tail Class Incremental Learning [20.267257778779992]
This paper introduces a two-stage framework designed to enhance long-tail class incremental learning.
We address the challenge posed by the under-representation of tail classes in long-tail class incremental learning.
The proposed framework can seamlessly integrate as a module with any class incremental learning method.
arXiv Detail & Related papers (2023-11-02T13:28:53Z) - Complementary Learning Subnetworks for Parameter-Efficient
Class-Incremental Learning [40.13416912075668]
We propose a rehearsal-free CIL approach that learns continually via the synergy between two Complementary Learning Subnetworks.
Our method achieves competitive results against state-of-the-art methods, especially in accuracy gain, memory cost, training efficiency, and task-order.
arXiv Detail & Related papers (2023-06-21T01:43:25Z) - Propheter: Prophetic Teacher Guided Long-Tailed Distribution Learning [44.947984354108094]
We propose an innovative long-tailed learning paradigm that breaks the bottleneck by guiding the learning of deep networks with external prior knowledge.
The proposed prophetic paradigm acts as a promising solution to the challenge of limited class knowledge in long-tailed datasets.
arXiv Detail & Related papers (2023-04-09T02:02:19Z) - Dense Network Expansion for Class Incremental Learning [61.00081795200547]
State-of-the-art approaches use a dynamic architecture based on network expansion (NE), in which a task expert is added per task.
A new NE method, dense network expansion (DNE), is proposed to achieve a better trade-off between accuracy and model complexity.
It outperforms the previous SOTA methods by a margin of 4% in terms of accuracy, with similar or even smaller model scale.
arXiv Detail & Related papers (2023-03-22T16:42:26Z) - Few-Shot Class-Incremental Learning by Sampling Multi-Phase Tasks [59.12108527904171]
A model should recognize new classes and maintain discriminability over old classes.
The task of recognizing few-shot new classes without forgetting old classes is called few-shot class-incremental learning (FSCIL)
We propose a new paradigm for FSCIL based on meta-learning by LearnIng Multi-phase Incremental Tasks (LIMIT)
arXiv Detail & Related papers (2022-03-31T13:46:41Z) - GROWN: GRow Only When Necessary for Continual Learning [39.56829374809613]
Catastrophic forgetting is a notorious issue in deep learning, referring to the fact that Deep Neural Networks (DNN) could forget the knowledge about earlier tasks when learning new tasks.
To address this issue, continual learning has been developed to learn new tasks sequentially and perform knowledge transfer from the old tasks to the new ones without forgetting.
GROWN is a novel end-to-end continual learning framework to dynamically grow the model only when necessary.
arXiv Detail & Related papers (2021-10-03T02:31:04Z) - Efficient Feature Transformations for Discriminative and Generative
Continual Learning [98.10425163678082]
We propose a simple task-specific feature map transformation strategy for continual learning.
Theses provide powerful flexibility for learning new tasks, achieved with minimal parameters added to the base architecture.
We demonstrate the efficacy and efficiency of our method with an extensive set of experiments in discriminative (CIFAR-100 and ImageNet-1K) and generative sequences of tasks.
arXiv Detail & Related papers (2021-03-25T01:48:14Z) - iTAML: An Incremental Task-Agnostic Meta-learning Approach [123.10294801296926]
Humans can continuously learn new knowledge as their experience grows.
Previous learning in deep neural networks can quickly fade out when they are trained on a new task.
We introduce a novel meta-learning approach that seeks to maintain an equilibrium between all encountered tasks.
arXiv Detail & Related papers (2020-03-25T21:42:48Z) - Adversarial Continual Learning [99.56738010842301]
We propose a hybrid continual learning framework that learns a disjoint representation for task-invariant and task-specific features.
Our model combines architecture growth to prevent forgetting of task-specific skills and an experience replay approach to preserve shared skills.
arXiv Detail & Related papers (2020-03-21T02:08:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.