Anomaly Detection in Video Data Based on Probabilistic Latent Space
Models
- URL: http://arxiv.org/abs/2003.07623v1
- Date: Tue, 17 Mar 2020 10:32:22 GMT
- Title: Anomaly Detection in Video Data Based on Probabilistic Latent Space
Models
- Authors: Giulia Slavic, Damian Campo, Mohamad Baydoun, Pablo Marin, David
Martin, Lucio Marcenaro, Carlo Regazzoni
- Abstract summary: A Variational Autoencoder (VAE) is used for reducing the dimensionality of video frames.
An Adapted Markov Jump Particle Filter defined by discrete and continuous inference levels is employed to predict the following frames.
Our method is evaluated on different video scenarios where a semi-autonomous vehicle performs a set of tasks in a closed environment.
- Score: 7.269230232703388
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper proposes a method for detecting anomalies in video data. A
Variational Autoencoder (VAE) is used for reducing the dimensionality of video
frames, generating latent space information that is comparable to
low-dimensional sensory data (e.g., positioning, steering angle), making
feasible the development of a consistent multi-modal architecture for
autonomous vehicles. An Adapted Markov Jump Particle Filter defined by discrete
and continuous inference levels is employed to predict the following frames and
detecting anomalies in new video sequences. Our method is evaluated on
different video scenarios where a semi-autonomous vehicle performs a set of
tasks in a closed environment.
Related papers
- Weakly Supervised Video Anomaly Detection and Localization with Spatio-Temporal Prompts [57.01985221057047]
This paper introduces a novel method that learnstemporal prompt embeddings for weakly supervised video anomaly detection and localization (WSVADL) based on pre-trained vision-language models (VLMs)
Our method achieves state-of-theart performance on three public benchmarks for the WSVADL task.
arXiv Detail & Related papers (2024-08-12T03:31:29Z) - VANE-Bench: Video Anomaly Evaluation Benchmark for Conversational LMMs [64.60035916955837]
VANE-Bench is a benchmark designed to assess the proficiency of Video-LMMs in detecting anomalies and inconsistencies in videos.
Our dataset comprises an array of videos synthetically generated using existing state-of-the-art text-to-video generation models.
We evaluate nine existing Video-LMMs, both open and closed sources, on this benchmarking task and find that most of the models encounter difficulties in effectively identifying the subtle anomalies.
arXiv Detail & Related papers (2024-06-14T17:59:01Z) - Dynamic Erasing Network Based on Multi-Scale Temporal Features for
Weakly Supervised Video Anomaly Detection [103.92970668001277]
We propose a Dynamic Erasing Network (DE-Net) for weakly supervised video anomaly detection.
We first propose a multi-scale temporal modeling module, capable of extracting features from segments of varying lengths.
Then, we design a dynamic erasing strategy, which dynamically assesses the completeness of the detected anomalies.
arXiv Detail & Related papers (2023-12-04T09:40:11Z) - Open-Vocabulary Video Anomaly Detection [57.552523669351636]
Video anomaly detection (VAD) with weak supervision has achieved remarkable performance in utilizing video-level labels to discriminate whether a video frame is normal or abnormal.
Recent studies attempt to tackle a more realistic setting, open-set VAD, which aims to detect unseen anomalies given seen anomalies and normal videos.
This paper takes a step further and explores open-vocabulary video anomaly detection (OVVAD), in which we aim to leverage pre-trained large models to detect and categorize seen and unseen anomalies.
arXiv Detail & Related papers (2023-11-13T02:54:17Z) - Point Cloud Video Anomaly Detection Based on Point Spatio-Temporal
Auto-Encoder [1.4340883856076097]
We propose Point Spatio-Temporal Auto-Encoder (PSTAE), an autoencoder framework that uses point cloud videos as input to detect anomalies in point cloud videos.
Our method sets a new state-of-the-art (SOTA) on the TIMo dataset.
arXiv Detail & Related papers (2023-06-04T10:30:28Z) - DAE : Discriminatory Auto-Encoder for multivariate time-series anomaly
detection in air transportation [68.8204255655161]
We propose a novel anomaly detection model called Discriminatory Auto-Encoder (DAE)
It uses the baseline of a regular LSTM-based auto-encoder but with several decoders, each getting data of a specific flight phase.
Results show that the DAE achieves better results in both accuracy and speed of detection.
arXiv Detail & Related papers (2021-09-08T14:07:55Z) - Robust Unsupervised Video Anomaly Detection by Multi-Path Frame
Prediction [61.17654438176999]
We propose a novel and robust unsupervised video anomaly detection method by frame prediction with proper design.
Our proposed method obtains the frame-level AUROC score of 88.3% on the CUHK Avenue dataset.
arXiv Detail & Related papers (2020-11-05T11:34:12Z) - Dynamic Bayesian Approach for decision-making in Ego-Things [8.577234269009042]
This paper presents a novel approach to detect abnormalities in dynamic systems based on multisensory data and feature selection.
Growing neural gas (GNG) is employed for clustering multisensory data into a set of nodes.
Our method uses a Markov Jump particle filter (MJPF) for state estimation and abnormality detection.
arXiv Detail & Related papers (2020-10-28T11:38:51Z) - Anomalous Motion Detection on Highway Using Deep Learning [14.617786106427834]
This paper presents a new anomaly detection dataset - the Highway Traffic Anomaly (HTA) dataset.
We evaluate state-of-the-art deep learning anomaly detection models and propose novel variations to these methods.
arXiv Detail & Related papers (2020-06-15T05:40:11Z) - Continual Learning of Predictive Models in Video Sequences via
Variational Autoencoders [6.698751933050415]
This paper proposes a method for performing continual learning of predictive models that facilitate the inference of future frames in video sequences.
An initial Variational Autoencoder, together with a set of fully connected neural networks are utilized to respectively learn the appearance of video frames and their dynamics at the latent space level.
arXiv Detail & Related papers (2020-06-02T21:17:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.