Incremental Embedding Learning via Zero-Shot Translation
- URL: http://arxiv.org/abs/2012.15497v1
- Date: Thu, 31 Dec 2020 08:21:37 GMT
- Title: Incremental Embedding Learning via Zero-Shot Translation
- Authors: Kun Wei, Cheng Deng, Xu Yang, and Maosen Li
- Abstract summary: Current state-of-the-art incremental learning methods tackle catastrophic forgetting problem in traditional classification networks.
We propose a novel class-incremental method for embedding network, named as zero-shot translation class-incremental method (ZSTCI)
In addition, ZSTCI can easily be combined with existing regularization-based incremental learning methods to further improve performance of embedding networks.
- Score: 65.94349068508863
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Modern deep learning methods have achieved great success in machine learning
and computer vision fields by learning a set of pre-defined datasets. Howerver,
these methods perform unsatisfactorily when applied into real-world situations.
The reason of this phenomenon is that learning new tasks leads the trained
model quickly forget the knowledge of old tasks, which is referred to as
catastrophic forgetting. Current state-of-the-art incremental learning methods
tackle catastrophic forgetting problem in traditional classification networks
and ignore the problem existing in embedding networks, which are the basic
networks for image retrieval, face recognition, zero-shot learning, etc.
Different from traditional incremental classification networks, the semantic
gap between the embedding spaces of two adjacent tasks is the main challenge
for embedding networks under incremental learning setting. Thus, we propose a
novel class-incremental method for embedding network, named as zero-shot
translation class-incremental method (ZSTCI), which leverages zero-shot
translation to estimate and compensate the semantic gap without any exemplars.
Then, we try to learn a unified representation for two adjacent tasks in
sequential learning process, which captures the relationships of previous
classes and current classes precisely. In addition, ZSTCI can easily be
combined with existing regularization-based incremental learning methods to
further improve performance of embedding networks. We conduct extensive
experiments on CUB-200-2011 and CIFAR100, and the experiment results prove the
effectiveness of our method. The code of our method has been released.
Related papers
- Bayesian Learning-driven Prototypical Contrastive Loss for Class-Incremental Learning [42.14439854721613]
We propose a prototypical network with a Bayesian learning-driven contrastive loss (BLCL) tailored specifically for class-incremental learning scenarios.
Our approach dynamically adapts the balance between the cross-entropy and contrastive loss functions with a Bayesian learning technique.
arXiv Detail & Related papers (2024-05-17T19:49:02Z) - Enhancing Visual Continual Learning with Language-Guided Supervision [76.38481740848434]
Continual learning aims to empower models to learn new tasks without forgetting previously acquired knowledge.
We argue that the scarce semantic information conveyed by the one-hot labels hampers the effective knowledge transfer across tasks.
Specifically, we use PLMs to generate semantic targets for each class, which are frozen and serve as supervision signals.
arXiv Detail & Related papers (2024-03-24T12:41:58Z) - Negotiated Representations to Prevent Forgetting in Machine Learning
Applications [0.0]
Catastrophic forgetting is a significant challenge in the field of machine learning.
We propose a novel method for preventing catastrophic forgetting in machine learning applications.
arXiv Detail & Related papers (2023-11-30T22:43:50Z) - Dense Network Expansion for Class Incremental Learning [61.00081795200547]
State-of-the-art approaches use a dynamic architecture based on network expansion (NE), in which a task expert is added per task.
A new NE method, dense network expansion (DNE), is proposed to achieve a better trade-off between accuracy and model complexity.
It outperforms the previous SOTA methods by a margin of 4% in terms of accuracy, with similar or even smaller model scale.
arXiv Detail & Related papers (2023-03-22T16:42:26Z) - Prototypical quadruplet for few-shot class incremental learning [24.814045065163135]
We propose a novel method that improves classification robustness by identifying a better embedding space using an improved contrasting loss.
Our approach retains previously acquired knowledge in the embedding space, even when trained with new classes.
We demonstrate the effectiveness of our method by showing that the embedding space remains intact after training the model with new classes and outperforms existing state-of-the-art algorithms in terms of accuracy across different sessions.
arXiv Detail & Related papers (2022-11-05T17:19:14Z) - On Generalizing Beyond Domains in Cross-Domain Continual Learning [91.56748415975683]
Deep neural networks often suffer from catastrophic forgetting of previously learned knowledge after learning a new task.
Our proposed approach learns new tasks under domain shift with accuracy boosts up to 10% on challenging datasets such as DomainNet and OfficeHome.
arXiv Detail & Related papers (2022-03-08T09:57:48Z) - Incremental Learning via Rate Reduction [26.323357617265163]
Current deep learning architectures suffer from catastrophic forgetting, a failure to retain knowledge of previously learned classes when incrementally trained on new classes.
We propose utilizing an alternative "white box" architecture derived from the principle of rate reduction, where each layer of the network is explicitly computed without back propagation.
Under this paradigm, we demonstrate that, given a pre-trained network and new data classes, our approach can provably construct a new network that emulates joint training with all past and new classes.
arXiv Detail & Related papers (2020-11-30T07:23:55Z) - Few-Shot Class-Incremental Learning [68.75462849428196]
We focus on a challenging but practical few-shot class-incremental learning (FSCIL) problem.
FSCIL requires CNN models to incrementally learn new classes from very few labelled samples, without forgetting the previously learned ones.
We represent the knowledge using a neural gas (NG) network, which can learn and preserve the topology of the feature manifold formed by different classes.
arXiv Detail & Related papers (2020-04-23T03:38:33Z) - Semantic Drift Compensation for Class-Incremental Learning [48.749630494026086]
Class-incremental learning of deep networks sequentially increases the number of classes to be classified.
We propose a new method to estimate the drift, called semantic drift, of features and compensate for it without the need of any exemplars.
arXiv Detail & Related papers (2020-04-01T13:31:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.