Neural Collapse Inspired Feature-Classifier Alignment for Few-Shot Class
Incremental Learning
- URL: http://arxiv.org/abs/2302.03004v1
- Date: Mon, 6 Feb 2023 18:39:40 GMT
- Title: Neural Collapse Inspired Feature-Classifier Alignment for Few-Shot Class
Incremental Learning
- Authors: Yibo Yang, Haobo Yuan, Xiangtai Li, Zhouchen Lin, Philip Torr, Dacheng
Tao
- Abstract summary: Few-shot class-incremental learning (FSCIL) has been a challenging problem as only a few training samples are accessible for each novel class in the new sessions.
We deal with this misalignment dilemma in FSCIL inspired by the recently discovered phenomenon named neural collapse.
We propose a neural collapse inspired framework for FSCIL. Experiments on the miniImageNet, CUB-200, and CIFAR-100 datasets demonstrate that our proposed framework outperforms the state-of-the-art performances.
- Score: 120.53458753007851
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Few-shot class-incremental learning (FSCIL) has been a challenging problem as
only a few training samples are accessible for each novel class in the new
sessions. Finetuning the backbone or adjusting the classifier prototypes
trained in the prior sessions would inevitably cause a misalignment between the
feature and classifier of old classes, which explains the well-known
catastrophic forgetting problem. In this paper, we deal with this misalignment
dilemma in FSCIL inspired by the recently discovered phenomenon named neural
collapse, which reveals that the last-layer features of the same class will
collapse into a vertex, and the vertices of all classes are aligned with the
classifier prototypes, which are formed as a simplex equiangular tight frame
(ETF). It corresponds to an optimal geometric structure for classification due
to the maximized Fisher Discriminant Ratio. We propose a neural collapse
inspired framework for FSCIL. A group of classifier prototypes are pre-assigned
as a simplex ETF for the whole label space, including the base session and all
the incremental sessions. During training, the classifier prototypes are not
learnable, and we adopt a novel loss function that drives the features into
their corresponding prototypes. Theoretical analysis shows that our method
holds the neural collapse optimality and does not break the feature-classifier
alignment in an incremental fashion. Experiments on the miniImageNet, CUB-200,
and CIFAR-100 datasets demonstrate that our proposed framework outperforms the
state-of-the-art performances. Code address:
https://github.com/NeuralCollapseApplications/FSCIL
Related papers
- Inductive Graph Few-shot Class Incremental Learning [34.19083477893245]
We introduce inductive GFSCIL that continually learns novel classes with newly emerging nodes.
Compared to the transductive GFSCIL, the inductive setting exacerbates catastrophic forgetting due to inaccessible previous data.
We propose a novel method, called Topology-based class Augmentation and Prototype calibration.
arXiv Detail & Related papers (2024-11-11T00:06:20Z) - Neural Collapse for Cross-entropy Class-Imbalanced Learning with Unconstrained ReLU Feature Model [25.61363481391964]
We show that when the training dataset is class-imbalanced, some Neural Collapse (NC) properties will no longer be true.
In this paper, we generalize NC to imbalanced regime for cross-entropy loss under the unconstrained ReLU feature model.
We find that the weights are aligned to the scaled and centered class-means with scaling factors depend on the number of training samples of each class.
arXiv Detail & Related papers (2024-01-04T04:53:31Z) - Few-Shot Class-Incremental Learning via Training-Free Prototype
Calibration [67.69532794049445]
We find a tendency for existing methods to misclassify the samples of new classes into base classes, which leads to the poor performance of new classes.
We propose a simple yet effective Training-frEE calibratioN (TEEN) strategy to enhance the discriminability of new classes.
arXiv Detail & Related papers (2023-12-08T18:24:08Z) - No Fear of Classifier Biases: Neural Collapse Inspired Federated
Learning with Synthetic and Fixed Classifier [10.491645205483051]
We propose a solution to the FL's classifier bias problem by utilizing a synthetic and fixed ETF classifier during training.
We devise several effective modules to better adapt the ETF structure in FL, achieving both high generalization and personalization.
Our method achieves state-of-the-art performances on CIFAR-10, CIFAR-100, and Tiny-ImageNet.
arXiv Detail & Related papers (2023-03-17T15:38:39Z) - Inducing Neural Collapse to a Fixed Hierarchy-Aware Frame for Reducing
Mistake Severity [0.0]
We propose to fix the linear classifier of a deep neural network to a Hierarchy-Aware Frame (HAFrame)
We demonstrate that our approach reduces the mistake severity of the model's predictions while maintaining its top-1 accuracy on several datasets.
arXiv Detail & Related papers (2023-03-10T03:44:01Z) - Understanding Imbalanced Semantic Segmentation Through Neural Collapse [81.89121711426951]
We show that semantic segmentation naturally brings contextual correlation and imbalanced distribution among classes.
We introduce a regularizer on feature centers to encourage the network to learn features closer to the appealing structure.
Our method ranks 1st and sets a new record on the ScanNet200 test leaderboard.
arXiv Detail & Related papers (2023-01-03T13:51:51Z) - Prediction Calibration for Generalized Few-shot Semantic Segmentation [101.69940565204816]
Generalized Few-shot Semantic (GFSS) aims to segment each image pixel into either base classes with abundant training examples or novel classes with only a handful of (e.g., 1-5) training images per class.
We build a cross-attention module that guides the classifier's final prediction using the fused multi-level features.
Our PCN outperforms the state-the-art alternatives by large margins.
arXiv Detail & Related papers (2022-10-15T13:30:12Z) - Do We Really Need a Learnable Classifier at the End of Deep Neural
Network? [118.18554882199676]
We study the potential of learning a neural network for classification with the classifier randomly as an ETF and fixed during training.
Our experimental results show that our method is able to achieve similar performances on image classification for balanced datasets.
arXiv Detail & Related papers (2022-03-17T04:34:28Z) - Self-Supervised Class Incremental Learning [51.62542103481908]
Existing Class Incremental Learning (CIL) methods are based on a supervised classification framework sensitive to data labels.
When updating them based on the new class data, they suffer from catastrophic forgetting: the model cannot discern old class data clearly from the new.
In this paper, we explore the performance of Self-Supervised representation learning in Class Incremental Learning (SSCIL) for the first time.
arXiv Detail & Related papers (2021-11-18T06:58:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.