Summarizing Stream Data for Memory-Constrained Online Continual Learning
- URL: http://arxiv.org/abs/2305.16645v2
- Date: Tue, 9 Jan 2024 06:16:01 GMT
- Title: Summarizing Stream Data for Memory-Constrained Online Continual Learning
- Authors: Jianyang Gu, Kai Wang, Wei Jiang, Yang You
- Abstract summary: We propose to Summarize the knowledge from the Stream Data (SSD) into more informative samples by distilling the training characteristics of real images.
We demonstrate that with limited extra computational overhead, SSD provides more than 3% accuracy boost for sequential CIFAR-100 under extremely restricted memory buffer.
- Score: 17.40956484727636
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Replay-based methods have proved their effectiveness on online continual
learning by rehearsing past samples from an auxiliary memory. With many efforts
made on improving training schemes based on the memory, however, the
information carried by each sample in the memory remains under-investigated.
Under circumstances with restricted storage space, the informativeness of the
memory becomes critical for effective replay. Although some works design
specific strategies to select representative samples, by only employing a small
number of original images, the storage space is still not well utilized. To
this end, we propose to Summarize the knowledge from the Stream Data (SSD) into
more informative samples by distilling the training characteristics of real
images. Through maintaining the consistency of training gradients and
relationship to the past tasks, the summarized samples are more representative
for the stream data compared to the original images. Extensive experiments are
conducted on multiple online continual learning benchmarks to support that the
proposed SSD method significantly enhances the replay effects. We demonstrate
that with limited extra computational overhead, SSD provides more than 3%
accuracy boost for sequential CIFAR-100 under extremely restricted memory
buffer. Code in https://github.com/vimar-gu/SSD.
Related papers
- Enhancing Consistency and Mitigating Bias: A Data Replay Approach for
Incremental Learning [100.7407460674153]
Deep learning systems are prone to catastrophic forgetting when learning from a sequence of tasks.
To mitigate the problem, a line of methods propose to replay the data of experienced tasks when learning new tasks.
However, it is not expected in practice considering the memory constraint or data privacy issue.
As a replacement, data-free data replay methods are proposed by inverting samples from the classification model.
arXiv Detail & Related papers (2024-01-12T12:51:12Z) - Improving Image Recognition by Retrieving from Web-Scale Image-Text Data [68.63453336523318]
We introduce an attention-based memory module, which learns the importance of each retrieved example from the memory.
Compared to existing approaches, our method removes the influence of the irrelevant retrieved examples, and retains those that are beneficial to the input query.
We show that it achieves state-of-the-art accuracies in ImageNet-LT, Places-LT and Webvision datasets.
arXiv Detail & Related papers (2023-04-11T12:12:05Z) - A Memory Transformer Network for Incremental Learning [64.0410375349852]
We study class-incremental learning, a training setup in which new classes of data are observed over time for the model to learn from.
Despite the straightforward problem formulation, the naive application of classification models to class-incremental learning results in the "catastrophic forgetting" of previously seen classes.
One of the most successful existing methods has been the use of a memory of exemplars, which overcomes the issue of catastrophic forgetting by saving a subset of past data into a memory bank and utilizing it to prevent forgetting when training future tasks.
arXiv Detail & Related papers (2022-10-10T08:27:28Z) - Improving Task-free Continual Learning by Distributionally Robust Memory
Evolution [9.345559196495746]
Task-free continual learning aims to learn a non-stationary data stream without explicit task definitions and not forget previous knowledge.
Existing methods overlook the high uncertainty in the memory data distribution.
We propose a principled memory evolution framework to dynamically evolve the memory data distribution.
arXiv Detail & Related papers (2022-07-15T02:16:09Z) - Sample Condensation in Online Continual Learning [13.041782266237]
Online Continual learning is a challenging learning scenario where the model must learn from a non-stationary stream of data.
We propose OLCGM, a novel replay-based continual learning strategy that uses knowledge condensation techniques to continuously compress the memory.
arXiv Detail & Related papers (2022-06-23T17:23:42Z) - Memory Replay with Data Compression for Continual Learning [80.95444077825852]
We propose memory replay with data compression to reduce the storage cost of old training samples.
We extensively validate this across several benchmarks of class-incremental learning and in a realistic scenario of object detection for autonomous driving.
arXiv Detail & Related papers (2022-02-14T10:26:23Z) - ACAE-REMIND for Online Continual Learning with Compressed Feature Replay [47.73014647702813]
We propose an auxiliary classifier auto-encoder (ACAE) module for feature replay at intermediate layers with high compression rates.
The reduced memory footprint per image allows us to save more exemplars for replay.
In our experiments, we conduct task-agnostic evaluation under online continual learning setting.
arXiv Detail & Related papers (2021-05-18T15:27:51Z) - Rainbow Memory: Continual Learning with a Memory of Diverse Samples [14.520337285540148]
We argue the importance of diversity of samples in an episodic memory.
We propose a novel memory management strategy based on per-sample classification uncertainty and data augmentation.
We show that the proposed method significantly improves the accuracy in blurry continual learning setups.
arXiv Detail & Related papers (2021-03-31T17:28:29Z) - Improving Computational Efficiency in Visual Reinforcement Learning via
Stored Embeddings [89.63764845984076]
We present Stored Embeddings for Efficient Reinforcement Learning (SEER)
SEER is a simple modification of existing off-policy deep reinforcement learning methods.
We show that SEER does not degrade the performance of RLizable agents while significantly saving computation and memory.
arXiv Detail & Related papers (2021-03-04T08:14:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.