Essentials for Class Incremental Learning
- URL: http://arxiv.org/abs/2102.09517v1
- Date: Thu, 18 Feb 2021 18:01:06 GMT
- Title: Essentials for Class Incremental Learning
- Authors: Sudhanshu Mittal and Silvio Galesso and Thomas Brox
- Abstract summary: Class-incremental learning results on CIFAR-100 and ImageNet improve over the state-of-the-art by a large margin, while keeping the approach simple.
- Score: 43.306374557919646
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Contemporary neural networks are limited in their ability to learn from
evolving streams of training data. When trained sequentially on new or evolving
tasks, their accuracy drops sharply, making them unsuitable for many real-world
applications. In this work, we shed light on the causes of this well-known yet
unsolved phenomenon - often referred to as catastrophic forgetting - in a
class-incremental setup. We show that a combination of simple components and a
loss that balances intra-task and inter-task learning can already resolve
forgetting to the same extent as more complex measures proposed in literature.
Moreover, we identify poor quality of the learned representation as another
reason for catastrophic forgetting in class-IL. We show that performance is
correlated with secondary class information (dark knowledge) learned by the
model and it can be improved by an appropriate regularizer. With these lessons
learned, class-incremental learning results on CIFAR-100 and ImageNet improve
over the state-of-the-art by a large margin, while keeping the approach simple.
Related papers
- CLOSER: Towards Better Representation Learning for Few-Shot Class-Incremental Learning [52.63674911541416]
Few-shot class-incremental learning (FSCIL) faces several challenges, such as overfitting and forgetting.
Our primary focus is representation learning on base classes to tackle the unique challenge of FSCIL.
We find that trying to secure the spread of features within a more confined feature space enables the learned representation to strike a better balance between transferability and discriminability.
arXiv Detail & Related papers (2024-10-08T02:23:16Z) - Normalization and effective learning rates in reinforcement learning [52.59508428613934]
Normalization layers have recently experienced a renaissance in the deep reinforcement learning and continual learning literature.
We show that normalization brings with it a subtle but important side effect: an equivalence between growth in the norm of the network parameters and decay in the effective learning rate.
We propose to make the learning rate schedule explicit with a simple re- parameterization which we call Normalize-and-Project.
arXiv Detail & Related papers (2024-07-01T20:58:01Z) - Reset It and Forget It: Relearning Last-Layer Weights Improves Continual and Transfer Learning [2.270857464465579]
This work identifies a simple pre-training mechanism that leads to representations exhibiting better continual and transfer learning.
The repeated resetting of weights in the last layer, which we nickname "zapping," was originally designed for a meta-continual-learning procedure.
We show it is surprisingly applicable in many settings beyond both meta-learning and continual learning.
arXiv Detail & Related papers (2023-10-12T02:52:14Z) - Class-Incremental Learning: A Survey [84.30083092434938]
Class-Incremental Learning (CIL) enables the learner to incorporate the knowledge of new classes incrementally.
CIL tends to catastrophically forget the characteristics of former ones, and its performance drastically degrades.
We provide a rigorous and unified evaluation of 17 methods in benchmark image classification tasks to find out the characteristics of different algorithms.
arXiv Detail & Related papers (2023-02-07T17:59:05Z) - Incremental Embedding Learning via Zero-Shot Translation [65.94349068508863]
Current state-of-the-art incremental learning methods tackle catastrophic forgetting problem in traditional classification networks.
We propose a novel class-incremental method for embedding network, named as zero-shot translation class-incremental method (ZSTCI)
In addition, ZSTCI can easily be combined with existing regularization-based incremental learning methods to further improve performance of embedding networks.
arXiv Detail & Related papers (2020-12-31T08:21:37Z) - Class-incremental Learning with Rectified Feature-Graph Preservation [24.098892115785066]
A central theme of this paper is to learn new classes that arrive in sequential phases over time.
We propose a weighted-Euclidean regularization for old knowledge preservation.
We show how it can work with binary cross-entropy to increase class separation for effective learning of new classes.
arXiv Detail & Related papers (2020-12-15T07:26:04Z) - Self-Supervised Learning Aided Class-Incremental Lifelong Learning [17.151579393716958]
We study the issue of catastrophic forgetting in class-incremental learning (Class-IL)
In training procedure of Class-IL, as the model has no knowledge about following tasks, it would only extract features necessary for tasks learned so far, whose information is insufficient for joint classification.
We propose to combine self-supervised learning, which can provide effective representations without requiring labels, with Class-IL to partly get around this problem.
arXiv Detail & Related papers (2020-06-10T15:15:27Z) - Few-Shot Class-Incremental Learning [68.75462849428196]
We focus on a challenging but practical few-shot class-incremental learning (FSCIL) problem.
FSCIL requires CNN models to incrementally learn new classes from very few labelled samples, without forgetting the previously learned ones.
We represent the knowledge using a neural gas (NG) network, which can learn and preserve the topology of the feature manifold formed by different classes.
arXiv Detail & Related papers (2020-04-23T03:38:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.