Improving Feature Generalizability with Multitask Learning in Class
Incremental Learning
- URL: http://arxiv.org/abs/2204.12915v1
- Date: Tue, 26 Apr 2022 07:47:54 GMT
- Title: Improving Feature Generalizability with Multitask Learning in Class
Incremental Learning
- Authors: Dong Ma, Chi Ian Tang, Cecilia Mascolo
- Abstract summary: Many deep learning applications, like keyword spotting, require the incorporation of new concepts (classes) over time, referred to as Class Incremental Learning (CIL)
The major challenge in CIL is catastrophic forgetting, i.e., preserving as much of the old knowledge as possible while learning new tasks.
We propose multitask learning during base model training to improve the feature generalizability.
Our approach enhances the average incremental learning accuracy by up to 5.5%, which enables more reliable and accurate keyword spotting over time.
- Score: 12.632121107536843
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Many deep learning applications, like keyword spotting, require the
incorporation of new concepts (classes) over time, referred to as Class
Incremental Learning (CIL). The major challenge in CIL is catastrophic
forgetting, i.e., preserving as much of the old knowledge as possible while
learning new tasks. Various techniques, such as regularization, knowledge
distillation, and the use of exemplars, have been proposed to resolve this
issue. However, prior works primarily focus on the incremental learning step,
while ignoring the optimization during the base model training. We hypothesize
that a more transferable and generalizable feature representation from the base
model would be beneficial to incremental learning.
In this work, we adopt multitask learning during base model training to
improve the feature generalizability. Specifically, instead of training a
single model with all the base classes, we decompose the base classes into
multiple subsets and regard each of them as a task. These tasks are trained
concurrently and a shared feature extractor is obtained for incremental
learning. We evaluate our approach on two datasets under various
configurations. The results show that our approach enhances the average
incremental learning accuracy by up to 5.5%, which enables more reliable and
accurate keyword spotting over time. Moreover, the proposed approach can be
combined with many existing techniques and provides additional performance
gain.
Related papers
- Achieving More with Less: A Tensor-Optimization-Powered Ensemble Method [53.170053108447455]
Ensemble learning is a method that leverages weak learners to produce a strong learner.
We design a smooth and convex objective function that leverages the concept of margin, making the strong learner more discriminative.
We then compare our algorithm with random forests of ten times the size and other classical methods across numerous datasets.
arXiv Detail & Related papers (2024-08-06T03:42:38Z) - Enhancing Visual Continual Learning with Language-Guided Supervision [76.38481740848434]
Continual learning aims to empower models to learn new tasks without forgetting previously acquired knowledge.
We argue that the scarce semantic information conveyed by the one-hot labels hampers the effective knowledge transfer across tasks.
Specifically, we use PLMs to generate semantic targets for each class, which are frozen and serve as supervision signals.
arXiv Detail & Related papers (2024-03-24T12:41:58Z) - Class incremental learning with probability dampening and cascaded gated classifier [4.285597067389559]
We propose a novel incremental regularisation approach called Margin Dampening and Cascaded Scaling.
The first combines a soft constraint and a knowledge distillation approach to preserve past knowledge while allowing forgetting new patterns.
We empirically show that our approach performs well on multiple benchmarks well-established baselines.
arXiv Detail & Related papers (2024-02-02T09:33:07Z) - Multi-View Class Incremental Learning [57.14644913531313]
Multi-view learning (MVL) has gained great success in integrating information from multiple perspectives of a dataset to improve downstream task performance.
This paper investigates a novel paradigm called multi-view class incremental learning (MVCIL), where a single model incrementally classifies new classes from a continual stream of views.
arXiv Detail & Related papers (2023-06-16T08:13:41Z) - Cross-Class Feature Augmentation for Class Incremental Learning [45.91253737682168]
We propose a novel class incremental learning approach by incorporating a feature augmentation technique motivated by adversarial attacks.
The proposed approach has a unique perspective to utilize the previous knowledge in class incremental learning since it augments features of arbitrary target classes.
Our method consistently outperforms existing class incremental learning methods by significant margins in various scenarios.
arXiv Detail & Related papers (2023-04-04T15:48:09Z) - Few-Shot Incremental Learning with Continually Evolved Classifiers [46.278573301326276]
Few-shot class-incremental learning (FSCIL) aims to design machine learning algorithms that can continually learn new concepts from a few data points.
The difficulty lies in that limited data from new classes not only lead to significant overfitting issues but also exacerbate the notorious catastrophic forgetting problems.
We propose a Continually Evolved CIF ( CEC) that employs a graph model to propagate context information between classifiers for adaptation.
arXiv Detail & Related papers (2021-04-07T10:54:51Z) - Meta-learning the Learning Trends Shared Across Tasks [123.10294801296926]
Gradient-based meta-learning algorithms excel at quick adaptation to new tasks with limited data.
Existing meta-learning approaches only depend on the current task information during the adaptation.
We propose a 'Path-aware' model-agnostic meta-learning approach.
arXiv Detail & Related papers (2020-10-19T08:06:47Z) - Adversarial Continual Learning [99.56738010842301]
We propose a hybrid continual learning framework that learns a disjoint representation for task-invariant and task-specific features.
Our model combines architecture growth to prevent forgetting of task-specific skills and an experience replay approach to preserve shared skills.
arXiv Detail & Related papers (2020-03-21T02:08:17Z) - XtarNet: Learning to Extract Task-Adaptive Representation for
Incremental Few-Shot Learning [24.144499302568565]
We propose XtarNet, which learns to extract task-adaptive representation (TAR) for facilitating incremental few-shot learning.
The TAR contains effective information for classifying both novel and base categories.
XtarNet achieves state-of-the-art incremental few-shot learning performance.
arXiv Detail & Related papers (2020-03-19T04:02:44Z) - Cognitively-Inspired Model for Incremental Learning Using a Few Examples [11.193504036335503]
Incremental learning attempts to develop a classifier which learns continuously from a stream of data segregated into different classes.
Deep learning approaches suffer from catastrophic forgetting when learning classes incrementally, while most incremental learning approaches require a large amount of training data per class.
We propose a novel approach inspired by the concept learning model of the hippocampus and the neocortex that represents each image class as centroids.
arXiv Detail & Related papers (2020-02-27T19:52:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.