Few-Shot Class-Incremental Learning via Entropy-Regularized Data-Free
Replay
- URL: http://arxiv.org/abs/2207.11213v1
- Date: Fri, 22 Jul 2022 17:30:51 GMT
- Title: Few-Shot Class-Incremental Learning via Entropy-Regularized Data-Free
Replay
- Authors: Huan Liu, Li Gu, Zhixiang Chi, Yang Wang, Yuanhao Yu, Jun Chen and Jin
Tang
- Abstract summary: Few-shot class-incremental learning (FSCIL) has been proposed aiming to enable a deep learning system to incrementally learn new classes with limited data.
We show through empirical results that adopting the data replay is surprisingly favorable.
We propose using data-free replay that can synthesize data by a generator without accessing real data.
- Score: 52.251188477192336
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Few-shot class-incremental learning (FSCIL) has been proposed aiming to
enable a deep learning system to incrementally learn new classes with limited
data. Recently, a pioneer claims that the commonly used replay-based method in
class-incremental learning (CIL) is ineffective and thus not preferred for
FSCIL. This has, if truth, a significant influence on the fields of FSCIL. In
this paper, we show through empirical results that adopting the data replay is
surprisingly favorable. However, storing and replaying old data can lead to a
privacy concern. To address this issue, we alternatively propose using
data-free replay that can synthesize data by a generator without accessing real
data. In observing the the effectiveness of uncertain data for knowledge
distillation, we impose entropy regularization in the generator training to
encourage more uncertain examples. Moreover, we propose to relabel the
generated data with one-hot-like labels. This modification allows the network
to learn by solely minimizing the cross-entropy loss, which mitigates the
problem of balancing different objectives in the conventional knowledge
distillation approach. Finally, we show extensive experimental results and
analysis on CIFAR-100, miniImageNet and CUB-200 to demonstrate the
effectiveness of our proposed one.
Related papers
- Reducing Catastrophic Forgetting in Online Class Incremental Learning Using Self-Distillation [3.8506666685467343]
In continual learning, previous knowledge is forgotten when a model learns new tasks.
In this paper, we tried to solve this problem by acquiring transferable knowledge through self-distillation.
Our proposed method outperformed conventional methods by experiments on CIFAR10, CIFAR100, and MiniimageNet datasets.
arXiv Detail & Related papers (2024-09-17T16:26:33Z) - Enabling On-Device Learning via Experience Replay with Efficient Dataset Condensation [15.915388740468815]
We propose an on-device framework that addresses the issue of identifying the most representative data to avoid significant information loss.
Specifically, to effectively handle unlabeled incoming data, we propose a pseudo-labeling technique designed for unlabeled on-device learning environments.
With a buffer capacity of just one sample per class, our method achieves an accuracy that outperforms the best existing baseline by 58.4%.
arXiv Detail & Related papers (2024-05-25T07:52:36Z) - Enhancing Consistency and Mitigating Bias: A Data Replay Approach for
Incremental Learning [100.7407460674153]
Deep learning systems are prone to catastrophic forgetting when learning from a sequence of tasks.
To mitigate the problem, a line of methods propose to replay the data of experienced tasks when learning new tasks.
However, it is not expected in practice considering the memory constraint or data privacy issue.
As a replacement, data-free data replay methods are proposed by inverting samples from the classification model.
arXiv Detail & Related papers (2024-01-12T12:51:12Z) - SURF: Semi-supervised Reward Learning with Data Augmentation for
Feedback-efficient Preference-based Reinforcement Learning [168.89470249446023]
We present SURF, a semi-supervised reward learning framework that utilizes a large amount of unlabeled samples with data augmentation.
In order to leverage unlabeled samples for reward learning, we infer pseudo-labels of the unlabeled samples based on the confidence of the preference predictor.
Our experiments demonstrate that our approach significantly improves the feedback-efficiency of the preference-based method on a variety of locomotion and robotic manipulation tasks.
arXiv Detail & Related papers (2022-03-18T16:50:38Z) - Always Be Dreaming: A New Approach for Data-Free Class-Incremental
Learning [73.24988226158497]
We consider the high-impact problem of Data-Free Class-Incremental Learning (DFCIL)
We propose a novel incremental distillation strategy for DFCIL, contributing a modified cross-entropy training and importance-weighted feature distillation.
Our method results in up to a 25.1% increase in final task accuracy (absolute difference) compared to SOTA DFCIL methods for common class-incremental benchmarks.
arXiv Detail & Related papers (2021-06-17T17:56:08Z) - Distilling Causal Effect of Data in Class-Incremental Learning [109.680987556265]
We propose a causal framework to explain the catastrophic forgetting in Class-Incremental Learning (CIL)
We derive a novel distillation method that is to mitigate to the existing anti-forgetting techniques, such as data replay and feature/label distillation.
arXiv Detail & Related papers (2021-03-02T14:14:10Z) - Class-incremental Learning with Rectified Feature-Graph Preservation [24.098892115785066]
A central theme of this paper is to learn new classes that arrive in sequential phases over time.
We propose a weighted-Euclidean regularization for old knowledge preservation.
We show how it can work with binary cross-entropy to increase class separation for effective learning of new classes.
arXiv Detail & Related papers (2020-12-15T07:26:04Z) - Provably Efficient Causal Reinforcement Learning with Confounded
Observational Data [135.64775986546505]
We study how to incorporate the dataset (observational data) collected offline, which is often abundantly available in practice, to improve the sample efficiency in the online setting.
We propose the deconfounded optimistic value iteration (DOVI) algorithm, which incorporates the confounded observational data in a provably efficient manner.
arXiv Detail & Related papers (2020-06-22T14:49:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.