Visual anomaly detection in video by variational autoencoder
- URL: http://arxiv.org/abs/2203.03872v1
- Date: Tue, 8 Mar 2022 06:22:04 GMT
- Title: Visual anomaly detection in video by variational autoencoder
- Authors: Faraz Waseem (yahoo), Rafael Perez Martinez (Stanford University),
Chris Wu (Stanford University)
- Abstract summary: An autoencoder is a neural network that is trained to recreate its input using latent representation of input also called a bottleneck layer.
In this paper we have demonstrated comparison between performance of convolutional LSTM versus a variation convolutional LSTM autoencoder.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Video anomalies detection is the intersection of anomaly detection and visual
intelligence. It has commercial applications in surveillance, security,
self-driving cars and crop monitoring. Videos can capture a variety of
anomalies. Due to efforts needed to label training data, unsupervised
approaches to train anomaly detection models for videos is more practical An
autoencoder is a neural network that is trained to recreate its input using
latent representation of input also called a bottleneck layer. Variational
autoencoder uses distribution (mean and variance) as compared to latent vector
as bottleneck layer and can have better regularization effect. In this paper we
have demonstrated comparison between performance of convolutional LSTM versus a
variation convolutional LSTM autoencoder
Related papers
- Abnormal Event Detection In Videos Using Deep Embedding [0.0]
Abnormal event detection or anomaly detection in surveillance videos is currently a challenge because of the diversity of possible events.
We propose an unsupervised approach for video anomaly detection with the aim to jointly optimize the objectives of the deep neural network.
arXiv Detail & Related papers (2024-09-15T17:44:51Z) - VANE-Bench: Video Anomaly Evaluation Benchmark for Conversational LMMs [64.60035916955837]
VANE-Bench is a benchmark designed to assess the proficiency of Video-LMMs in detecting anomalies and inconsistencies in videos.
Our dataset comprises an array of videos synthetically generated using existing state-of-the-art text-to-video generation models.
We evaluate nine existing Video-LMMs, both open and closed sources, on this benchmarking task and find that most of the models encounter difficulties in effectively identifying the subtle anomalies.
arXiv Detail & Related papers (2024-06-14T17:59:01Z) - Dynamic Erasing Network Based on Multi-Scale Temporal Features for
Weakly Supervised Video Anomaly Detection [103.92970668001277]
We propose a Dynamic Erasing Network (DE-Net) for weakly supervised video anomaly detection.
We first propose a multi-scale temporal modeling module, capable of extracting features from segments of varying lengths.
Then, we design a dynamic erasing strategy, which dynamically assesses the completeness of the detected anomalies.
arXiv Detail & Related papers (2023-12-04T09:40:11Z) - Video Anomaly Detection using GAN [0.0]
This thesis study aims to offer the solution for this use case so that human resources won't be required to keep an eye out for any unusual activity in the surveillance system records.
We have developed a novel generative adversarial network (GAN) based anomaly detection model.
arXiv Detail & Related papers (2023-11-23T16:41:30Z) - Self-Supervised Masked Convolutional Transformer Block for Anomaly
Detection [122.4894940892536]
We present a novel self-supervised masked convolutional transformer block (SSMCTB) that comprises the reconstruction-based functionality at a core architectural level.
In this work, we extend our previous self-supervised predictive convolutional attentive block (SSPCAB) with a 3D masked convolutional layer, a transformer for channel-wise attention, as well as a novel self-supervised objective based on Huber loss.
arXiv Detail & Related papers (2022-09-25T04:56:10Z) - Anomaly detection in surveillance videos using transformer based
attention model [3.2968779106235586]
This research suggests using a weakly supervised strategy to avoid annotating anomalous segments in training videos.
The proposed framework is validated on real-world dataset i.e. ShanghaiTech Campus dataset.
arXiv Detail & Related papers (2022-06-03T12:19:39Z) - Anomaly Detection in Multi-Agent Trajectories for Automated Driving [2.5211566369910967]
Similar to humans, automated vehicles are supposed to perform anomaly detection.
Our innovation is the ability to jointly learn multiple trajectories of a dynamic number of agents.
arXiv Detail & Related papers (2021-10-15T08:07:31Z) - DAE : Discriminatory Auto-Encoder for multivariate time-series anomaly
detection in air transportation [68.8204255655161]
We propose a novel anomaly detection model called Discriminatory Auto-Encoder (DAE)
It uses the baseline of a regular LSTM-based auto-encoder but with several decoders, each getting data of a specific flight phase.
Results show that the DAE achieves better results in both accuracy and speed of detection.
arXiv Detail & Related papers (2021-09-08T14:07:55Z) - Autoencoding Variational Autoencoder [56.05008520271406]
We study the implications of this behaviour on the learned representations and also the consequences of fixing it by introducing a notion of self consistency.
We show that encoders trained with our self-consistency approach lead to representations that are robust (insensitive) to perturbations in the input introduced by adversarial attacks.
arXiv Detail & Related papers (2020-12-07T14:16:14Z) - Robust Unsupervised Video Anomaly Detection by Multi-Path Frame
Prediction [61.17654438176999]
We propose a novel and robust unsupervised video anomaly detection method by frame prediction with proper design.
Our proposed method obtains the frame-level AUROC score of 88.3% on the CUHK Avenue dataset.
arXiv Detail & Related papers (2020-11-05T11:34:12Z) - Unsupervised Video Anomaly Detection via Normalizing Flows with Implicit
Latent Features [8.407188666535506]
Most existing methods use an autoencoder to learn to reconstruct normal videos.
We propose an implicit two-path AE (ITAE), a structure in which two encoders implicitly model appearance and motion features.
For the complex distribution of normal scenes, we suggest normal density estimation of ITAE features.
NF models intensify ITAE performance by learning normality through implicitly learned features.
arXiv Detail & Related papers (2020-10-15T05:02:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.