Adaptive Online Incremental Learning for Evolving Data Streams
- URL: http://arxiv.org/abs/2201.01633v1
- Date: Wed, 5 Jan 2022 14:25:53 GMT
- Title: Adaptive Online Incremental Learning for Evolving Data Streams
- Authors: Si-si Zhang, Jian-wei Liu, Xin Zuo
- Abstract summary: The first major difficulty is concept drift, that is, the probability distribution in the streaming data would change as the data arrives.
The second major difficulty is catastrophic forgetting, that is, forgetting what we have learned before when learning new knowledge.
Our research builds on this observation and attempts to overcome these difficulties.
- Score: 4.3386084277869505
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Recent years have witnessed growing interests in online incremental learning.
However, there are three major challenges in this area. The first major
difficulty is concept drift, that is, the probability distribution in the
streaming data would change as the data arrives. The second major difficulty is
catastrophic forgetting, that is, forgetting what we have learned before when
learning new knowledge. The last one we often ignore is the learning of the
latent representation. Only good latent representation can improve the
prediction accuracy of the model. Our research builds on this observation and
attempts to overcome these difficulties. To this end, we propose an Adaptive
Online Incremental Learning for evolving data streams (AOIL). We use
auto-encoder with the memory module, on the one hand, we obtained the latent
features of the input, on the other hand, according to the reconstruction loss
of the auto-encoder with memory module, we could successfully detect the
existence of concept drift and trigger the update mechanism, adjust the model
parameters in time. In addition, we divide features, which are derived from the
activation of the hidden layers, into two parts, which are used to extract the
common and private features respectively. By means of this approach, the model
could learn the private features of the new coming instances, but do not forget
what we have learned in the past (shared features), which reduces the
occurrence of catastrophic forgetting. At the same time, to get the fusion
feature vector we use the self-attention mechanism to effectively fuse the
extracted features, which further improved the latent representation learning.
Related papers
- T4P: Test-Time Training of Trajectory Prediction via Masked Autoencoder and Actor-specific Token Memory [39.021321011792786]
Trajectory prediction is a challenging problem that requires considering interactions among multiple actors.
Data-driven approaches have been used to address this complex problem, but they suffer from unreliable predictions under distribution shifts during test time.
We propose several online learning methods using regression loss from the ground truth of observed data.
Our method surpasses the performance of existing state-of-the-art online learning methods in terms of both prediction accuracy and computational efficiency.
arXiv Detail & Related papers (2024-03-15T06:47:14Z) - Class incremental learning with probability dampening and cascaded gated classifier [4.285597067389559]
We propose a novel incremental regularisation approach called Margin Dampening and Cascaded Scaling.
The first combines a soft constraint and a knowledge distillation approach to preserve past knowledge while allowing forgetting new patterns.
We empirically show that our approach performs well on multiple benchmarks well-established baselines.
arXiv Detail & Related papers (2024-02-02T09:33:07Z) - Towards Causal Deep Learning for Vulnerability Detection [31.59558109518435]
We introduce do calculus based causal learning to software engineering models.
Our results show that CausalVul consistently improved the model accuracy, robustness and OOD performance.
arXiv Detail & Related papers (2023-10-12T00:51:06Z) - Adaptive Cross Batch Normalization for Metric Learning [75.91093210956116]
Metric learning is a fundamental problem in computer vision.
We show that it is equally important to ensure that the accumulated embeddings are up to date.
In particular, it is necessary to circumvent the representational drift between the accumulated embeddings and the feature embeddings at the current training iteration.
arXiv Detail & Related papers (2023-03-30T03:22:52Z) - Confidence Attention and Generalization Enhanced Distillation for
Continuous Video Domain Adaptation [62.458968086881555]
Continuous Video Domain Adaptation (CVDA) is a scenario where a source model is required to adapt to a series of individually available changing target domains.
We propose a Confidence-Attentive network with geneRalization enhanced self-knowledge disTillation (CART) to address the challenge in CVDA.
arXiv Detail & Related papers (2023-03-18T16:40:10Z) - Online Deep Learning from Doubly-Streaming Data [17.119725174036653]
This paper investigates a new online learning problem with doubly-streaming data, where the data streams are described by feature spaces that constantly evolve.
A plausible idea to overcome the challenges is to establish relationship between the pre-and-post evolving feature spaces.
We propose a novel OLD3S paradigm, where a shared latent subspace is discovered to summarize information from the old and new feature spaces.
arXiv Detail & Related papers (2022-04-25T17:06:39Z) - MetAug: Contrastive Learning via Meta Feature Augmentation [28.708395209321846]
We argue that contrastive learning heavily relies on informative features, or "hard" (positive or negative) features.
The key challenge toward exploring such features is that the source multi-view data is generated by applying random data augmentations.
We propose to directly augment the features in latent space, thereby learning discriminative representations without a large amount of input data.
arXiv Detail & Related papers (2022-03-10T02:35:39Z) - Towards Open-World Feature Extrapolation: An Inductive Graph Learning
Approach [80.8446673089281]
We propose a new learning paradigm with graph representation and learning.
Our framework contains two modules: 1) a backbone network (e.g., feedforward neural nets) as a lower model takes features as input and outputs predicted labels; 2) a graph neural network as an upper model learns to extrapolate embeddings for new features via message passing over a feature-data graph built from observed data.
arXiv Detail & Related papers (2021-10-09T09:02:45Z) - Machine Unlearning of Features and Labels [72.81914952849334]
We propose first scenarios for unlearning and labels in machine learning models.
Our approach builds on the concept of influence functions and realizes unlearning through closed-form updates of model parameters.
arXiv Detail & Related papers (2021-08-26T04:42:24Z) - Reducing Representation Drift in Online Continual Learning [87.71558506591937]
We study the online continual learning paradigm, where agents must learn from a changing distribution with constrained memory and compute.
In this work we instead focus on the change in representations of previously observed data due to the introduction of previously unobserved class samples in the incoming data stream.
arXiv Detail & Related papers (2021-04-11T15:19:30Z) - Value-driven Hindsight Modelling [68.658900923595]
Value estimation is a critical component of the reinforcement learning (RL) paradigm.
Model learning can make use of the rich transition structure present in sequences of observations, but this approach is usually not sensitive to the reward function.
We develop an approach for representation learning in RL that sits in between these two extremes.
This provides tractable prediction targets that are directly relevant for a task, and can thus accelerate learning the value function.
arXiv Detail & Related papers (2020-02-19T18:10:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.