LARA: A Light and Anti-overfitting Retraining Approach for Unsupervised
Time Series Anomaly Detection
- URL: http://arxiv.org/abs/2310.05668v4
- Date: Fri, 23 Feb 2024 05:58:26 GMT
- Title: LARA: A Light and Anti-overfitting Retraining Approach for Unsupervised
Time Series Anomaly Detection
- Authors: Feiyi Chen, Zhen Qin, Yingying Zhang, Shuiguang Deng, Yi Xiao,
Guansong Pang and Qingsong Wen
- Abstract summary: We propose a Light and Anti-overfitting Retraining Approach (LARA) for deep variational auto-encoder based time series anomaly detection methods (VAEs)
This work aims to make three novel contributions: 1) the retraining process is formulated as a convex problem and can converge at a fast rate as well as prevent overfitting; 2) designing a ruminate block, which leverages the historical data without the need to store them; and 3) mathematically proving that when fine-tuning the latent vector and reconstructed data, the linear formations can achieve the least adjusting errors between the ground truths and the fine-tuned ones.
- Score: 49.52429991848581
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Most of current anomaly detection models assume that the normal pattern
remains same all the time. However, the normal patterns of Web services change
dramatically and frequently. The model trained on old-distribution data is
outdated after such changes. Retraining the whole model every time is
expensive. Besides, at the beginning of normal pattern changes, there is not
enough observation data from the new distribution. Retraining a large neural
network model with limited data is vulnerable to overfitting. Thus, we propose
a Light and Anti-overfitting Retraining Approach (LARA) for deep variational
auto-encoder based time series anomaly detection methods (VAEs). This work aims
to make three novel contributions: 1) the retraining process is formulated as a
convex problem and can converge at a fast rate as well as prevent overfitting;
2) designing a ruminate block, which leverages the historical data without the
need to store them; 3) mathematically proving that when fine-tuning the latent
vector and reconstructed data, the linear formations can achieve the least
adjusting errors between the ground truths and the fine-tuned ones.
Moreover, we have performed many experiments to verify that retraining LARA
with even 43 time slots of data from new distribution can result in its
competitive F1 Score in comparison with the state-of-the-art anomaly detection
models trained with sufficient data. Besides, we verify its light overhead.
Related papers
- Enhancing Consistency and Mitigating Bias: A Data Replay Approach for
Incremental Learning [100.7407460674153]
Deep learning systems are prone to catastrophic forgetting when learning from a sequence of tasks.
To mitigate the problem, a line of methods propose to replay the data of experienced tasks when learning new tasks.
However, it is not expected in practice considering the memory constraint or data privacy issue.
As a replacement, data-free data replay methods are proposed by inverting samples from the classification model.
arXiv Detail & Related papers (2024-01-12T12:51:12Z) - MAD: Self-Supervised Masked Anomaly Detection Task for Multivariate Time
Series [14.236092062538653]
Masked Anomaly Detection (MAD) is a general self-supervised learning task for multivariate time series anomaly detection.
By randomly masking a portion of the inputs and training a model to estimate them, MAD is an improvement over the traditional left-to-right next step prediction (NSP) task.
Our experimental results demonstrate that MAD can achieve better anomaly detection rates over traditional NSP approaches.
arXiv Detail & Related papers (2022-05-04T14:55:42Z) - Deep Generative model with Hierarchical Latent Factors for Time Series
Anomaly Detection [40.21502451136054]
This work presents DGHL, a new family of generative models for time series anomaly detection.
A top-down Convolution Network maps a novel hierarchical latent space to time series windows, exploiting temporal dynamics to encode information efficiently.
Our method outperformed current state-of-the-art models on four popular benchmark datasets.
arXiv Detail & Related papers (2022-02-15T17:19:44Z) - Monte Carlo EM for Deep Time Series Anomaly Detection [6.312089019297173]
Time series data are often corrupted by outliers or other kinds of anomalies.
Recent approaches to anomaly detection and forecasting assume that the proportion of anomalies in the training data is small enough to ignore.
We present a technique for augmenting existing time series models so that they explicitly account for anomalies in the training data.
arXiv Detail & Related papers (2021-12-29T07:52:36Z) - Deep Visual Anomaly detection with Negative Learning [18.79849041106952]
In this paper, we propose anomaly detection with negative learning (ADNL), which employs the negative learning concept for the enhancement of anomaly detection.
The idea is to limit the reconstruction capability of a generative model using the given a small amount of anomaly examples.
This way, the network not only learns to reconstruct normal data but also encloses the normal distribution far from the possible distribution of anomalies.
arXiv Detail & Related papers (2021-05-24T01:48:44Z) - CutPaste: Self-Supervised Learning for Anomaly Detection and
Localization [59.719925639875036]
We propose a framework for building anomaly detectors using normal training data only.
We first learn self-supervised deep representations and then build a generative one-class classifier on learned representations.
Our empirical study on MVTec anomaly detection dataset demonstrates the proposed algorithm is general to be able to detect various types of real-world defects.
arXiv Detail & Related papers (2021-04-08T19:04:55Z) - Anomaly Detection of Time Series with Smoothness-Inducing Sequential
Variational Auto-Encoder [59.69303945834122]
We present a Smoothness-Inducing Sequential Variational Auto-Encoder (SISVAE) model for robust estimation and anomaly detection of time series.
Our model parameterizes mean and variance for each time-stamp with flexible neural networks.
We show the effectiveness of our model on both synthetic datasets and public real-world benchmarks.
arXiv Detail & Related papers (2021-02-02T06:15:15Z) - NVAE-GAN Based Approach for Unsupervised Time Series Anomaly Detection [19.726089445453734]
Time series anomaly detection is a common but challenging task in many industries.
It is difficult to detect anomalies in time series with high accuracy, due to noisy data collected from real world.
We propose our anomaly detection model: Time series to Image VAE (T2IVAE)
arXiv Detail & Related papers (2021-01-08T08:35:15Z) - Variational Bayesian Unlearning [54.26984662139516]
We study the problem of approximately unlearning a Bayesian model from a small subset of the training data to be erased.
We show that it is equivalent to minimizing an evidence upper bound which trades off between fully unlearning from erased data vs. not entirely forgetting the posterior belief.
In model training with VI, only an approximate (instead of exact) posterior belief given the full data can be obtained, which makes unlearning even more challenging.
arXiv Detail & Related papers (2020-10-24T11:53:00Z) - TadGAN: Time Series Anomaly Detection Using Generative Adversarial
Networks [73.01104041298031]
TadGAN is an unsupervised anomaly detection approach built on Generative Adversarial Networks (GANs)
To capture the temporal correlations of time series, we use LSTM Recurrent Neural Networks as base models for Generators and Critics.
To demonstrate the performance and generalizability of our approach, we test several anomaly scoring techniques and report the best-suited one.
arXiv Detail & Related papers (2020-09-16T15:52:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.