Continual Prototype Evolution: Learning Online from Non-Stationary Data
Streams
- URL: http://arxiv.org/abs/2009.00919v4
- Date: Tue, 6 Apr 2021 10:40:42 GMT
- Title: Continual Prototype Evolution: Learning Online from Non-Stationary Data
Streams
- Authors: Matthias De Lange, Tinne Tuytelaars
- Abstract summary: We introduce a system to enable learning and prediction at any point in time.
In contrast to the major body of work in continual learning, data streams are processed in an online fashion.
We obtain state-of-the-art performance by a significant margin on eight benchmarks, including three highly imbalanced data streams.
- Score: 42.525141660788
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Attaining prototypical features to represent class distributions is well
established in representation learning. However, learning prototypes online
from streaming data proves a challenging endeavor as they rapidly become
outdated, caused by an ever-changing parameter space during the learning
process. Additionally, continual learning does not assume the data stream to be
stationary, typically resulting in catastrophic forgetting of previous
knowledge. As a first, we introduce a system addressing both problems, where
prototypes evolve continually in a shared latent space, enabling learning and
prediction at any point in time. In contrast to the major body of work in
continual learning, data streams are processed in an online fashion, without
additional task-information, and an efficient memory scheme provides robustness
to imbalanced data streams. Besides nearest neighbor based prediction, learning
is facilitated by a novel objective function, encouraging cluster density about
the class prototype and increased inter-class variance. Furthermore, the latent
space quality is elevated by pseudo-prototypes in each batch, constituted by
replay of exemplars from memory. As an additional contribution, we generalize
the existing paradigms in continual learning to incorporate data incremental
learning from data streams by formalizing a two-agent learner-evaluator
framework. We obtain state-of-the-art performance by a significant margin on
eight benchmarks, including three highly imbalanced data streams.
Related papers
- RPS: A Generic Reservoir Patterns Sampler [1.09784964592609]
We introduce an approach that harnesses a weighted reservoir to facilitate direct pattern sampling from streaming batch data.
We present a generic algorithm capable of addressing temporal biases and handling various pattern types, including sequential, weighted, and unweighted itemsets.
arXiv Detail & Related papers (2024-10-31T16:25:21Z) - Premonition: Using Generative Models to Preempt Future Data Changes in
Continual Learning [63.850451635362425]
Continual learning requires a model to adapt to ongoing changes in the data distribution.
We show that the combination of a large language model and an image generation model can similarly provide useful premonitions.
We find that the backbone of our pre-trained networks can learn representations useful for the downstream continual learning problem.
arXiv Detail & Related papers (2024-03-12T06:29:54Z) - Random Representations Outperform Online Continually Learned Representations [68.42776779425978]
We show that existing online continually trained deep networks produce inferior representations compared to a simple pre-defined random transforms.
Our method, called RanDumb, significantly outperforms state-of-the-art continually learned representations across all online continual learning benchmarks.
Our study reveals the significant limitations of representation learning, particularly in low-exemplar and online continual learning scenarios.
arXiv Detail & Related papers (2024-02-13T22:07:29Z) - CTP: Towards Vision-Language Continual Pretraining via Compatible
Momentum Contrast and Topology Preservation [128.00940554196976]
Vision-Language Continual Pretraining (VLCP) has shown impressive results on diverse downstream tasks by offline training on large-scale datasets.
To support the study of Vision-Language Continual Pretraining (VLCP), we first contribute a comprehensive and unified benchmark dataset P9D.
The data from each industry as an independent task supports continual learning and conforms to the real-world long-tail nature to simulate pretraining on web data.
arXiv Detail & Related papers (2023-08-14T13:53:18Z) - Prototype-Sample Relation Distillation: Towards Replay-Free Continual
Learning [14.462797749666992]
We propose a holistic approach to jointly learn the representation and class prototypes.
We propose a novel distillation loss that constrains class prototypes to maintain relative similarities as compared to new task data.
This method yields state-of-the-art performance in the task-incremental setting.
arXiv Detail & Related papers (2023-03-26T16:35:45Z) - Continual Learning with Optimal Transport based Mixture Model [17.398605698033656]
We propose an online mixture model learning approach based on nice properties of the mature optimal transport theory (OT-MM)
Our proposed method can significantly outperform the current state-of-the-art baselines.
arXiv Detail & Related papers (2022-11-30T06:40:29Z) - Data augmentation on-the-fly and active learning in data stream
classification [9.367903089535684]
There is an emerging need for predictive models to be trained on-the-fly.
Learning models have access to more labelled data without the need to increase the active learning budget.
Augmented Queues significantly improves the performance in terms of learning quality and speed.
arXiv Detail & Related papers (2022-10-13T09:57:08Z) - A Memory Transformer Network for Incremental Learning [64.0410375349852]
We study class-incremental learning, a training setup in which new classes of data are observed over time for the model to learn from.
Despite the straightforward problem formulation, the naive application of classification models to class-incremental learning results in the "catastrophic forgetting" of previously seen classes.
One of the most successful existing methods has been the use of a memory of exemplars, which overcomes the issue of catastrophic forgetting by saving a subset of past data into a memory bank and utilizing it to prevent forgetting when training future tasks.
arXiv Detail & Related papers (2022-10-10T08:27:28Z) - Nonstationary data stream classification with online active learning and
siamese neural networks [11.501721946030779]
An emerging need for online learning methods that train predictive models on-the-fly.
A series of open challenges, however, hinder their deployment in practice.
We propose the ActiSiamese algorithm, which addresses these challenges by combining online active learning, siamese networks, and a multi-queue memory.
arXiv Detail & Related papers (2022-10-03T17:16:03Z) - Online Continual Learning with Natural Distribution Shifts: An Empirical
Study with Visual Data [101.6195176510611]
"Online" continual learning enables evaluating both information retention and online learning efficacy.
In online continual learning, each incoming small batch of data is first used for testing and then added to the training set, making the problem truly online.
We introduce a new benchmark for online continual visual learning that exhibits large scale and natural distribution shifts.
arXiv Detail & Related papers (2021-08-20T06:17:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.