Unsupervised Online Anomaly Detection On Irregularly Sampled Or Missing
Valued Time-Series Data Using LSTM Networks
- URL: http://arxiv.org/abs/2005.12005v1
- Date: Mon, 25 May 2020 09:41:04 GMT
- Title: Unsupervised Online Anomaly Detection On Irregularly Sampled Or Missing
Valued Time-Series Data Using LSTM Networks
- Authors: Oguzhan Karaahmetoglu (1 and 2), Fatih Ilhan (1 and 2), Ismail Balaban
(2), Suleyman Serdar Kozat (1 and 2) ((1) Bilkent University, (2) DataBoss
A.S.)
- Abstract summary: We study anomaly detection and introduce an algorithm that processes variable length, irregularly sampled sequences or sequences with missing values.
Our algorithm is fully unsupervised, however, can be readily extended to supervised or semisupervised cases.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We study anomaly detection and introduce an algorithm that processes variable
length, irregularly sampled sequences or sequences with missing values. Our
algorithm is fully unsupervised, however, can be readily extended to supervised
or semisupervised cases when the anomaly labels are present as remarked
throughout the paper. Our approach uses the Long Short Term Memory (LSTM)
networks in order to extract temporal features and find the most relevant
feature vectors for anomaly detection. We incorporate the sampling time
information to our model by modulating the standard LSTM model with time
modulation gates. After obtaining the most relevant features from the LSTM, we
label the sequences using a Support Vector Data Descriptor (SVDD) model. We
introduce a loss function and then jointly optimize the feature extraction and
sequence processing mechanisms in an end-to-end manner. Through this joint
optimization, the LSTM extracts the most relevant features for anomaly
detection later to be used in the SVDD, hence completely removes the need for
feature selection by expert knowledge. Furthermore, we provide a training
algorithm for the online setup, where we optimize our model parameters with
individual sequences as the new data arrives. Finally, on real-life datasets,
we show that our model significantly outperforms the standard approaches thanks
to its combination of LSTM with SVDD and joint optimization.
Related papers
- Anomaly Detection of Tabular Data Using LLMs [54.470648484612866]
We show that pre-trained large language models (LLMs) are zero-shot batch-level anomaly detectors.
We propose an end-to-end fine-tuning strategy to bring out the potential of LLMs in detecting real anomalies.
arXiv Detail & Related papers (2024-06-24T04:17:03Z) - Improved AutoEncoder with LSTM module and KL divergence [3.1168862003127797]
We propose Improved AutoEncoder with LSTM module and Kullback-Leibler divergence (IAE-LSTM-KL) model in this paper.
The efficacy of the IAE-LSTM-KL model is validated through experiments on both synthetic and real-world datasets.
arXiv Detail & Related papers (2024-04-30T04:11:21Z) - Low-Rank Representations Meets Deep Unfolding: A Generalized and
Interpretable Network for Hyperspectral Anomaly Detection [41.50904949744355]
Current hyperspectral anomaly detection (HAD) benchmark datasets suffer from low resolution, simple background, and small size of the detection data.
These factors also limit the performance of the well-known low-rank representation (LRR) models in terms of robustness.
We build a new set of HAD benchmark datasets for improving the robustness of the HAD algorithm in complex scenarios, AIR-HAD for short.
arXiv Detail & Related papers (2024-02-23T14:15:58Z) - Graph Spatiotemporal Process for Multivariate Time Series Anomaly
Detection with Missing Values [67.76168547245237]
We introduce a novel framework called GST-Pro, which utilizes a graphtemporal process and anomaly scorer to detect anomalies.
Our experimental results show that the GST-Pro method can effectively detect anomalies in time series data and outperforms state-of-the-art methods.
arXiv Detail & Related papers (2024-01-11T10:10:16Z) - Embedded feature selection in LSTM networks with multi-objective
evolutionary ensemble learning for time series forecasting [49.1574468325115]
We present a novel feature selection method embedded in Long Short-Term Memory networks.
Our approach optimize the weights and biases of the LSTM in a partitioned manner.
Experimental evaluations on air quality time series data from Italy and southeast Spain demonstrate that our method substantially improves the ability generalization of conventional LSTMs.
arXiv Detail & Related papers (2023-12-29T08:42:10Z) - No Imputation Needed: A Switch Approach to Irregularly Sampled Time Series [41.387374646018344]
We present SLAN (Switch LSTM Aggregate Network), which utilizes a group of LSTMs to model irregularly-sampled time series (ISTS) without imputation.
SLAN exploits the irregularity information to explicitly capture each sensor's local summary and maintains a global summary state throughout the observational period.
We demonstrate the efficacy of SLAN on two public datasets, namely, MIMIC-III, and Physionet 2012.
arXiv Detail & Related papers (2023-09-15T18:43:41Z) - A Bi-LSTM Autoencoder Framework for Anomaly Detection -- A Case Study of
a Wind Power Dataset [2.094022863940315]
Anomalies refer to data points or events that deviate from normal and homogeneous events.
This study presents a novel framework for time series anomaly detection using a combination of Bi-LSTM architecture and Autoencoder.
The Bi-LSTM Autoencoder model achieved a classification accuracy of 96.79% and outperformed more commonly used LSTM Autoencoder models.
arXiv Detail & Related papers (2023-03-17T00:24:28Z) - Imputing Missing Observations with Time Sliced Synthetic Minority
Oversampling Technique [0.3973560285628012]
We present a simple yet novel time series imputation technique with the goal of constructing an irregular time series that is uniform across every sample in a data set.
We fix a grid defined by the midpoints of non-overlapping bins (dubbed "slices") of observation times and ensure that each sample has values for all of the features at that given time.
This allows one to both impute fully missing observations to allow uniform time series classification across the entire data and, in special cases, to impute individually missing features.
arXiv Detail & Related papers (2022-01-14T19:23:24Z) - Adaptive Memory Networks with Self-supervised Learning for Unsupervised
Anomaly Detection [54.76993389109327]
Unsupervised anomaly detection aims to build models to detect unseen anomalies by only training on the normal data.
We propose a novel approach called Adaptive Memory Network with Self-supervised Learning (AMSL) to address these challenges.
AMSL incorporates a self-supervised learning module to learn general normal patterns and an adaptive memory fusion module to learn rich feature representations.
arXiv Detail & Related papers (2022-01-03T03:40:21Z) - Anomaly Detection of Time Series with Smoothness-Inducing Sequential
Variational Auto-Encoder [59.69303945834122]
We present a Smoothness-Inducing Sequential Variational Auto-Encoder (SISVAE) model for robust estimation and anomaly detection of time series.
Our model parameterizes mean and variance for each time-stamp with flexible neural networks.
We show the effectiveness of our model on both synthetic datasets and public real-world benchmarks.
arXiv Detail & Related papers (2021-02-02T06:15:15Z) - Contextual-Bandit Anomaly Detection for IoT Data in Distributed
Hierarchical Edge Computing [65.78881372074983]
IoT devices can hardly afford complex deep neural networks (DNN) models, and offloading anomaly detection tasks to the cloud incurs long delay.
We propose and build a demo for an adaptive anomaly detection approach for distributed hierarchical edge computing (HEC) systems.
We show that our proposed approach significantly reduces detection delay without sacrificing accuracy, as compared to offloading detection tasks to the cloud.
arXiv Detail & Related papers (2020-04-15T06:13:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.