Self-Supervised Class Incremental Learning
- URL: http://arxiv.org/abs/2111.11208v1
- Date: Thu, 18 Nov 2021 06:58:19 GMT
- Title: Self-Supervised Class Incremental Learning
- Authors: Zixuan Ni, Siliang Tang, Yueting Zhuang
- Abstract summary: Existing Class Incremental Learning (CIL) methods are based on a supervised classification framework sensitive to data labels.
When updating them based on the new class data, they suffer from catastrophic forgetting: the model cannot discern old class data clearly from the new.
In this paper, we explore the performance of Self-Supervised representation learning in Class Incremental Learning (SSCIL) for the first time.
- Score: 51.62542103481908
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Existing Class Incremental Learning (CIL) methods are based on a supervised
classification framework sensitive to data labels. When updating them based on
the new class data, they suffer from catastrophic forgetting: the model cannot
discern old class data clearly from the new. In this paper, we explore the
performance of Self-Supervised representation learning in Class Incremental
Learning (SSCIL) for the first time, which discards data labels and the model's
classifiers. To comprehensively discuss the difference in performance between
supervised and self-supervised methods in CIL, we set up three different class
incremental schemes: Random Class Scheme, Semantic Class Scheme, and Cluster
Scheme, to simulate various class incremental learning scenarios. Besides, we
propose Linear Evaluation Protocol (LEP) and Generalization Evaluation Protocol
(GEP) to metric the model's representation classification ability and
generalization in CIL. Our experiments (on ImageNet-100 and ImageNet) show that
SSCIL has better anti-forgetting ability and robustness than supervised
strategies in CIL. To understand what alleviates the catastrophic forgetting in
SSCIL, we study the major components of SSCIL and conclude that (1) the
composition of different data augmentation improves the quality of the model's
representation and the \textit{Grayscale} operation reduces the system noise of
data augmentation in SSCIL. (2) the projector, like a buffer, reduces
unnecessary parameter updates of the model in SSCIL and increases the
robustness of the model. Although the performance of SSCIL is significantly
higher than supervised methods in CIL, there is still an apparent gap with
joint learning. Our exploration gives a baseline of self-supervised class
incremental learning on large-scale datasets and contributes some forward
strategies for mitigating the catastrophic forgetting in CIL.
Related papers
- Class-Imbalanced Semi-Supervised Learning for Large-Scale Point Cloud
Semantic Segmentation via Decoupling Optimization [64.36097398869774]
Semi-supervised learning (SSL) has been an active research topic for large-scale 3D scene understanding.
The existing SSL-based methods suffer from severe training bias due to class imbalance and long-tail distributions of the point cloud data.
We introduce a new decoupling optimization framework, which disentangles feature representation learning and classifier in an alternative optimization manner to shift the bias decision boundary effectively.
arXiv Detail & Related papers (2024-01-13T04:16:40Z) - PILoRA: Prototype Guided Incremental LoRA for Federated Class-Incremental Learning [41.984652077669104]
Experimental results on standard datasets indicate that our method outperforms the state-of-the-art approaches significantly.
Our method exhibits strong robustness and superiority in different settings and degrees of data heterogeneity.
arXiv Detail & Related papers (2024-01-04T06:46:19Z) - Class-Incremental Learning: A Survey [84.30083092434938]
Class-Incremental Learning (CIL) enables the learner to incorporate the knowledge of new classes incrementally.
CIL tends to catastrophically forget the characteristics of former ones, and its performance drastically degrades.
We provide a rigorous and unified evaluation of 17 methods in benchmark image classification tasks to find out the characteristics of different algorithms.
arXiv Detail & Related papers (2023-02-07T17:59:05Z) - Learning from Mistakes: Self-Regularizing Hierarchical Representations
in Point Cloud Semantic Segmentation [15.353256018248103]
LiDAR semantic segmentation has gained attention to accomplish fine-grained scene understanding.
We present a coarse-to-fine setup that LEArns from classification mistaKes (LEAK) derived from a standard model.
Our LEAK approach is very general and can be seamlessly applied on top of any segmentation architecture.
arXiv Detail & Related papers (2023-01-26T14:52:30Z) - Mitigating Forgetting in Online Continual Learning via Contrasting
Semantically Distinct Augmentations [22.289830907729705]
Online continual learning (OCL) aims to enable model learning from a non-stationary data stream to continuously acquire new knowledge as well as retain the learnt one.
Main challenge comes from the "catastrophic forgetting" issue -- the inability to well remember the learnt knowledge while learning the new ones.
arXiv Detail & Related papers (2022-11-10T05:29:43Z) - Class-Incremental Learning with Strong Pre-trained Models [97.84755144148535]
Class-incremental learning (CIL) has been widely studied under the setting of starting from a small number of classes (base classes)
We explore an understudied real-world setting of CIL that starts with a strong model pre-trained on a large number of base classes.
Our proposed method is robust and generalizes to all analyzed CIL settings.
arXiv Detail & Related papers (2022-04-07T17:58:07Z) - No Fear of Heterogeneity: Classifier Calibration for Federated Learning
with Non-IID Data [78.69828864672978]
A central challenge in training classification models in the real-world federated system is learning with non-IID data.
We propose a novel and simple algorithm called Virtual Representations (CCVR), which adjusts the classifier using virtual representations sampled from an approximated ssian mixture model.
Experimental results demonstrate that CCVR state-of-the-art performance on popular federated learning benchmarks including CIFAR-10, CIFAR-100, and CINIC-10.
arXiv Detail & Related papers (2021-06-09T12:02:29Z) - Few-Shot Incremental Learning with Continually Evolved Classifiers [46.278573301326276]
Few-shot class-incremental learning (FSCIL) aims to design machine learning algorithms that can continually learn new concepts from a few data points.
The difficulty lies in that limited data from new classes not only lead to significant overfitting issues but also exacerbate the notorious catastrophic forgetting problems.
We propose a Continually Evolved CIF ( CEC) that employs a graph model to propagate context information between classifiers for adaptation.
arXiv Detail & Related papers (2021-04-07T10:54:51Z) - Class-incremental Learning with Rectified Feature-Graph Preservation [24.098892115785066]
A central theme of this paper is to learn new classes that arrive in sequential phases over time.
We propose a weighted-Euclidean regularization for old knowledge preservation.
We show how it can work with binary cross-entropy to increase class separation for effective learning of new classes.
arXiv Detail & Related papers (2020-12-15T07:26:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.