Tracking changes using Kullback-Leibler divergence for the continual
learning
- URL: http://arxiv.org/abs/2210.04865v1
- Date: Mon, 10 Oct 2022 17:30:41 GMT
- Title: Tracking changes using Kullback-Leibler divergence for the continual
learning
- Authors: Sebasti\'an Basterrech and Michal Wo\'zniak
- Abstract summary: This article introduces a novel method for monitoring changes in the probabilistic distribution of multi-dimensional data streams.
As a measure of the rapidity of changes, we analyze the popular Kullback-Leibler divergence.
We show how to use this metric to predict the concept drift occurrence and understand its nature.
- Score: 2.0305676256390934
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recently, continual learning has received a lot of attention. One of the
significant problems is the occurrence of \emph{concept drift}, which consists
of changing probabilistic characteristics of the incoming data. In the case of
the classification task, this phenomenon destabilizes the model's performance
and negatively affects the achieved prediction quality. Most current methods
apply statistical learning and similarity analysis over the raw data. However,
similarity analysis in streaming data remains a complex problem due to time
limitation, non-precise values, fast decision speed, scalability, etc. This
article introduces a novel method for monitoring changes in the probabilistic
distribution of multi-dimensional data streams. As a measure of the rapidity of
changes, we analyze the popular Kullback-Leibler divergence. During the
experimental study, we show how to use this metric to predict the concept drift
occurrence and understand its nature. The obtained results encourage further
work on the proposed methods and its application in the real tasks where the
prediction of the future appearance of concept drift plays a crucial role, such
as predictive maintenance.
Related papers
- Dreaming Learning [41.94295877935867]
Introducing new information to a machine learning system can interfere with previously stored data.
We propose a training algorithm inspired by Stuart Kauffman's notion of the Adjacent Possible.
It predisposes the neural network to smoothly accept and integrate data sequences with different statistical characteristics than expected.
arXiv Detail & Related papers (2024-10-23T09:17:31Z) - Unsupervised Assessment of Landscape Shifts Based on Persistent Entropy and Topological Preservation [0.0]
A drift in the input data can have negative consequences on a learning predictor and the system's stability.
In this article, we introduce a novel framework for monitoring changes in multi-dimensional data streams.
The framework operates in both unsupervised and supervised environments.
arXiv Detail & Related papers (2024-10-05T14:57:52Z) - Causally-Aware Spatio-Temporal Multi-Graph Convolution Network for Accurate and Reliable Traffic Prediction [5.200012764049096]
This study focuses on an instance of--temporal problem--traffic prediction--to demonstrate an advanced deep learning model for making accurate and reliable forecast.
We propose an end-to-end traffic prediction framework that leverages three primary components to accurate and reliable traffic predictions.
Experimental results on two real-world traffic datasets demonstrate that the method outperforms several state-of-the-art models in prediction accuracy.
arXiv Detail & Related papers (2024-08-23T14:35:54Z) - Performative Time-Series Forecasting [71.18553214204978]
We formalize performative time-series forecasting (PeTS) from a machine-learning perspective.
We propose a novel approach, Feature Performative-Shifting (FPS), which leverages the concept of delayed response to anticipate distribution shifts.
We conduct comprehensive experiments using multiple time-series models on COVID-19 and traffic forecasting tasks.
arXiv Detail & Related papers (2023-10-09T18:34:29Z) - Discovering Predictable Latent Factors for Time Series Forecasting [39.08011991308137]
We develop a novel framework for inferring the intrinsic latent factors implied by the observable time series.
We introduce three characteristics, i.e., predictability, sufficiency, and identifiability, and model these characteristics via the powerful deep latent dynamics models.
Empirical results on multiple real datasets show the efficiency of our method for different kinds of time series forecasting.
arXiv Detail & Related papers (2023-03-18T14:37:37Z) - Adapting to Continuous Covariate Shift via Online Density Ratio Estimation [64.8027122329609]
Dealing with distribution shifts is one of the central challenges for modern machine learning.
We propose an online method that can appropriately reuse historical information.
Our density ratio estimation method is proven to perform well by enjoying a dynamic regret bound.
arXiv Detail & Related papers (2023-02-06T04:03:33Z) - TACTiS: Transformer-Attentional Copulas for Time Series [76.71406465526454]
estimation of time-varying quantities is a fundamental component of decision making in fields such as healthcare and finance.
We propose a versatile method that estimates joint distributions using an attention-based decoder.
We show that our model produces state-of-the-art predictions on several real-world datasets.
arXiv Detail & Related papers (2022-02-07T21:37:29Z) - Accurate and Robust Feature Importance Estimation under Distribution
Shifts [49.58991359544005]
PRoFILE is a novel feature importance estimation method.
We show significant improvements over state-of-the-art approaches, both in terms of fidelity and robustness.
arXiv Detail & Related papers (2020-09-30T05:29:01Z) - On Robustness and Transferability of Convolutional Neural Networks [147.71743081671508]
Modern deep convolutional networks (CNNs) are often criticized for not generalizing under distributional shifts.
We study the interplay between out-of-distribution and transfer performance of modern image classification CNNs for the first time.
We find that increasing both the training set and model sizes significantly improve the distributional shift robustness.
arXiv Detail & Related papers (2020-07-16T18:39:04Z) - New Perspectives on the Use of Online Learning for Congestion Level
Prediction over Traffic Data [6.664111208927475]
This work focuses on classification over time series data.
When a time series is generated by non-stationary phenomena, the pattern relating the series with the class to be predicted may evolve over time.
Online learning methods incrementally learn from new data samples arriving over time, and accommodate eventual changes along the data stream.
arXiv Detail & Related papers (2020-03-27T09:44:57Z) - Value-driven Hindsight Modelling [68.658900923595]
Value estimation is a critical component of the reinforcement learning (RL) paradigm.
Model learning can make use of the rich transition structure present in sequences of observations, but this approach is usually not sensitive to the reward function.
We develop an approach for representation learning in RL that sits in between these two extremes.
This provides tractable prediction targets that are directly relevant for a task, and can thus accelerate learning the value function.
arXiv Detail & Related papers (2020-02-19T18:10:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.