GROWN: GRow Only When Necessary for Continual Learning
- URL: http://arxiv.org/abs/2110.00908v1
- Date: Sun, 3 Oct 2021 02:31:04 GMT
- Title: GROWN: GRow Only When Necessary for Continual Learning
- Authors: Li Yang, Sen Lin, Junshan Zhang, Deliang Fan
- Abstract summary: Catastrophic forgetting is a notorious issue in deep learning, referring to the fact that Deep Neural Networks (DNN) could forget the knowledge about earlier tasks when learning new tasks.
To address this issue, continual learning has been developed to learn new tasks sequentially and perform knowledge transfer from the old tasks to the new ones without forgetting.
GROWN is a novel end-to-end continual learning framework to dynamically grow the model only when necessary.
- Score: 39.56829374809613
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Catastrophic forgetting is a notorious issue in deep learning, referring to
the fact that Deep Neural Networks (DNN) could forget the knowledge about
earlier tasks when learning new tasks. To address this issue, continual
learning has been developed to learn new tasks sequentially and perform
knowledge transfer from the old tasks to the new ones without forgetting. While
recent structure-based learning methods show the capability of alleviating the
forgetting problem, these methods start from a redundant full-size network and
require a complex learning process to gradually grow-and-prune or search the
network structure for each task, which is inefficient. To address this problem
and enable efficient network expansion for new tasks, we first develop a
learnable sparse growth method eliminating the additional pruning/searching
step in previous structure-based methods. Building on this learnable sparse
growth method, we then propose GROWN, a novel end-to-end continual learning
framework to dynamically grow the model only when necessary. Different from all
previous structure-based methods, GROWN starts from a small seed network,
instead of a full-sized one. We validate GROWN on multiple datasets against
state-of-the-art methods, which shows superior performance in both accuracy and
model size. For example, we achieve 1.0\% accuracy gain on average compared to
the current SOTA results on CIFAR-100 Superclass 20 tasks setting.
Related papers
- TaE: Task-aware Expandable Representation for Long Tail Class Incremental Learning [42.630413950957795]
We introduce a novel Task-aware Expandable (TaE) framework to learn diverse representations from each incremental task.
TaE achieves state-of-the-art performance.
arXiv Detail & Related papers (2024-02-08T16:37:04Z) - MixtureGrowth: Growing Neural Networks by Recombining Learned Parameters [19.358670728803336]
Most deep neural networks are trained under fixed network architectures and require retraining when the architecture changes.
To avoid this, one can grow from a small network by adding random weights over time to gradually achieve the target network size.
This naive approach falls short in practice as it brings too much noise to the growing process.
arXiv Detail & Related papers (2023-11-07T11:37:08Z) - Dense Network Expansion for Class Incremental Learning [61.00081795200547]
State-of-the-art approaches use a dynamic architecture based on network expansion (NE), in which a task expert is added per task.
A new NE method, dense network expansion (DNE), is proposed to achieve a better trade-off between accuracy and model complexity.
It outperforms the previous SOTA methods by a margin of 4% in terms of accuracy, with similar or even smaller model scale.
arXiv Detail & Related papers (2023-03-22T16:42:26Z) - Neural Weight Search for Scalable Task Incremental Learning [6.413209417643468]
Task incremental learning aims to enable a system to maintain its performance on previously learned tasks while learning new tasks, solving the problem of catastrophic forgetting.
One promising approach is to build an individual network or sub-network for future tasks.
This leads to an ever-growing memory due to saving extra weights for new tasks and how to address this issue has remained an open problem in task incremental learning.
arXiv Detail & Related papers (2022-11-24T23:30:23Z) - Rethinking Task-Incremental Learning Baselines [5.771817160915079]
We present a simple yet effective adjustment network (SAN) for task incremental learning that achieves near state-of-the-art performance.
We investigate this approach on both 3D point cloud object (ModelNet40) and 2D image (CIFAR10, CIFAR100, MiniImageNet, MNIST, PermutedMNIST, notMNIST, SVHN, and FashionMNIST) recognition tasks.
arXiv Detail & Related papers (2022-05-23T14:52:38Z) - Rectification-based Knowledge Retention for Continual Learning [49.1447478254131]
Deep learning models suffer from catastrophic forgetting when trained in an incremental learning setting.
We propose a novel approach to address the task incremental learning problem, which involves training a model on new tasks that arrive in an incremental manner.
Our approach can be used in both the zero-shot and non zero-shot task incremental learning settings.
arXiv Detail & Related papers (2021-03-30T18:11:30Z) - Incremental Embedding Learning via Zero-Shot Translation [65.94349068508863]
Current state-of-the-art incremental learning methods tackle catastrophic forgetting problem in traditional classification networks.
We propose a novel class-incremental method for embedding network, named as zero-shot translation class-incremental method (ZSTCI)
In addition, ZSTCI can easily be combined with existing regularization-based incremental learning methods to further improve performance of embedding networks.
arXiv Detail & Related papers (2020-12-31T08:21:37Z) - Semantic Drift Compensation for Class-Incremental Learning [48.749630494026086]
Class-incremental learning of deep networks sequentially increases the number of classes to be classified.
We propose a new method to estimate the drift, called semantic drift, of features and compensate for it without the need of any exemplars.
arXiv Detail & Related papers (2020-04-01T13:31:19Z) - iTAML: An Incremental Task-Agnostic Meta-learning Approach [123.10294801296926]
Humans can continuously learn new knowledge as their experience grows.
Previous learning in deep neural networks can quickly fade out when they are trained on a new task.
We introduce a novel meta-learning approach that seeks to maintain an equilibrium between all encountered tasks.
arXiv Detail & Related papers (2020-03-25T21:42:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.