Knowledge Transfer-Driven Few-Shot Class-Incremental Learning
- URL: http://arxiv.org/abs/2306.10942v1
- Date: Mon, 19 Jun 2023 14:02:45 GMT
- Title: Knowledge Transfer-Driven Few-Shot Class-Incremental Learning
- Authors: Ye Wang, Yaxiong Wang, Guoshuai Zhao, and Xueming Qian
- Abstract summary: Few-shot class-incremental learning (FSCIL) aims to continually learn new classes using a few samples while not forgetting the old classes.
Despite the advance of existing FSCIL methods, the proposed knowledge transfer learning schemes are sub-optimal due to the insufficient optimization for the model's plasticity.
We propose a Random Episode Sampling and Augmentation (RESA) strategy that relies on diverse pseudo incremental tasks as agents to achieve the knowledge transfer.
- Score: 23.163459923345556
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Few-shot class-incremental learning (FSCIL) aims to continually learn new
classes using a few samples while not forgetting the old classes. The key of
this task is effective knowledge transfer from the base session to the
incremental sessions. Despite the advance of existing FSCIL methods, the
proposed knowledge transfer learning schemes are sub-optimal due to the
insufficient optimization for the model's plasticity. To address this issue, we
propose a Random Episode Sampling and Augmentation (RESA) strategy that relies
on diverse pseudo incremental tasks as agents to achieve the knowledge
transfer. Concretely, RESA mimics the real incremental setting and constructs
pseudo incremental tasks globally and locally, where the global pseudo
incremental tasks are designed to coincide with the learning objective of FSCIL
and the local pseudo incremental tasks are designed to improve the model's
plasticity, respectively. Furthermore, to make convincing incremental
predictions, we introduce a complementary model with a squared
Euclidean-distance classifier as the auxiliary module, which couples with the
widely used cosine classifier to form our whole architecture. By such a way,
equipped with model decoupling strategy, we can maintain the model's stability
while enhancing the model's plasticity. Extensive quantitative and qualitative
experiments on three popular FSCIL benchmark datasets demonstrate that our
proposed method, named Knowledge Transfer-driven Relation Complementation
Network (KT-RCNet), outperforms almost all prior methods. More precisely, the
average accuracy of our proposed KT-RCNet outperforms the second-best method by
a margin of 5.26%, 3.49%, and 2.25% on miniImageNet, CIFAR100, and CUB200,
respectively. Our code is available at
https://github.com/YeZiLaiXi/KT-RCNet.git.
Related papers
- Knowledge Adaptation Network for Few-Shot Class-Incremental Learning [23.90555521006653]
Few-shot class-incremental learning aims to incrementally recognize new classes using a few samples.
One of the effective methods to solve this challenge is to construct prototypical evolution classifiers.
Because representations for new classes are weak and biased, we argue such a strategy is suboptimal.
arXiv Detail & Related papers (2024-09-18T07:51:38Z) - USER: Unified Semantic Enhancement with Momentum Contrast for Image-Text
Retrieval [115.28586222748478]
Image-Text Retrieval (ITR) aims at searching for the target instances that are semantically relevant to the given query from the other modality.
Existing approaches typically suffer from two major limitations.
arXiv Detail & Related papers (2023-01-17T12:42:58Z) - FOSTER: Feature Boosting and Compression for Class-Incremental Learning [52.603520403933985]
Deep neural networks suffer from catastrophic forgetting when learning new categories.
We propose a novel two-stage learning paradigm FOSTER, empowering the model to learn new categories adaptively.
arXiv Detail & Related papers (2022-04-10T11:38:33Z) - Class-Incremental Learning with Strong Pre-trained Models [97.84755144148535]
Class-incremental learning (CIL) has been widely studied under the setting of starting from a small number of classes (base classes)
We explore an understudied real-world setting of CIL that starts with a strong model pre-trained on a large number of base classes.
Our proposed method is robust and generalizes to all analyzed CIL settings.
arXiv Detail & Related papers (2022-04-07T17:58:07Z) - Self-Supervised Class Incremental Learning [51.62542103481908]
Existing Class Incremental Learning (CIL) methods are based on a supervised classification framework sensitive to data labels.
When updating them based on the new class data, they suffer from catastrophic forgetting: the model cannot discern old class data clearly from the new.
In this paper, we explore the performance of Self-Supervised representation learning in Class Incremental Learning (SSCIL) for the first time.
arXiv Detail & Related papers (2021-11-18T06:58:19Z) - Learning to Continuously Optimize Wireless Resource in a Dynamic
Environment: A Bilevel Optimization Perspective [52.497514255040514]
This work develops a new approach that enables data-driven methods to continuously learn and optimize resource allocation strategies in a dynamic environment.
We propose to build the notion of continual learning into wireless system design, so that the learning model can incrementally adapt to the new episodes.
Our design is based on a novel bilevel optimization formulation which ensures certain fairness" across different data samples.
arXiv Detail & Related papers (2021-05-03T07:23:39Z) - Few-Shot Incremental Learning with Continually Evolved Classifiers [46.278573301326276]
Few-shot class-incremental learning (FSCIL) aims to design machine learning algorithms that can continually learn new concepts from a few data points.
The difficulty lies in that limited data from new classes not only lead to significant overfitting issues but also exacerbate the notorious catastrophic forgetting problems.
We propose a Continually Evolved CIF ( CEC) that employs a graph model to propagate context information between classifiers for adaptation.
arXiv Detail & Related papers (2021-04-07T10:54:51Z) - Adaptive Consistency Regularization for Semi-Supervised Transfer
Learning [31.66745229673066]
We consider semi-supervised learning and transfer learning jointly, leading to a more practical and competitive paradigm.
To better exploit the value of both pre-trained weights and unlabeled target examples, we introduce adaptive consistency regularization.
Our proposed adaptive consistency regularization outperforms state-of-the-art semi-supervised learning techniques such as Pseudo Label, Mean Teacher, and MixMatch.
arXiv Detail & Related papers (2021-03-03T05:46:39Z) - Towards Accurate Knowledge Transfer via Target-awareness Representation
Disentanglement [56.40587594647692]
We propose a novel transfer learning algorithm, introducing the idea of Target-awareness REpresentation Disentanglement (TRED)
TRED disentangles the relevant knowledge with respect to the target task from the original source model and used as a regularizer during fine-tuning the target model.
Experiments on various real world datasets show that our method stably improves the standard fine-tuning by more than 2% in average.
arXiv Detail & Related papers (2020-10-16T17:45:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.