Anomaly detection in surveillance videos using transformer based
attention model
- URL: http://arxiv.org/abs/2206.01524v2
- Date: Mon, 6 Jun 2022 10:04:53 GMT
- Title: Anomaly detection in surveillance videos using transformer based
attention model
- Authors: Kapil Deshpande, Narinder Singh Punn, Sanjay Kumar Sonbhadra, Sonali
Agarwal
- Abstract summary: This research suggests using a weakly supervised strategy to avoid annotating anomalous segments in training videos.
The proposed framework is validated on real-world dataset i.e. ShanghaiTech Campus dataset.
- Score: 3.2968779106235586
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Surveillance footage can catch a wide range of realistic anomalies. This
research suggests using a weakly supervised strategy to avoid annotating
anomalous segments in training videos, which is time consuming. In this
approach only video level labels are used to obtain frame level anomaly scores.
Weakly supervised video anomaly detection (WSVAD) suffers from the wrong
identification of abnormal and normal instances during the training process.
Therefore it is important to extract better quality features from the available
videos. WIth this motivation, the present paper uses better quality
transformer-based features named Videoswin Features followed by the attention
layer based on dilated convolution and self attention to capture long and short
range dependencies in temporal domain. This gives us a better understanding of
available videos. The proposed framework is validated on real-world dataset
i.e. ShanghaiTech Campus dataset which results in competitive performance than
current state-of-the-art methods. The model and the code are available at
https://github.com/kapildeshpande/Anomaly-Detection-in-Surveillance-Videos
Related papers
- Weakly Supervised Video Anomaly Detection and Localization with Spatio-Temporal Prompts [57.01985221057047]
This paper introduces a novel method that learnstemporal prompt embeddings for weakly supervised video anomaly detection and localization (WSVADL) based on pre-trained vision-language models (VLMs)
Our method achieves state-of-theart performance on three public benchmarks for the WSVADL task.
arXiv Detail & Related papers (2024-08-12T03:31:29Z) - VANE-Bench: Video Anomaly Evaluation Benchmark for Conversational LMMs [64.60035916955837]
VANE-Bench is a benchmark designed to assess the proficiency of Video-LMMs in detecting anomalies and inconsistencies in videos.
Our dataset comprises an array of videos synthetically generated using existing state-of-the-art text-to-video generation models.
We evaluate nine existing Video-LMMs, both open and closed sources, on this benchmarking task and find that most of the models encounter difficulties in effectively identifying the subtle anomalies.
arXiv Detail & Related papers (2024-06-14T17:59:01Z) - Dynamic Erasing Network Based on Multi-Scale Temporal Features for
Weakly Supervised Video Anomaly Detection [103.92970668001277]
We propose a Dynamic Erasing Network (DE-Net) for weakly supervised video anomaly detection.
We first propose a multi-scale temporal modeling module, capable of extracting features from segments of varying lengths.
Then, we design a dynamic erasing strategy, which dynamically assesses the completeness of the detected anomalies.
arXiv Detail & Related papers (2023-12-04T09:40:11Z) - Towards Video Anomaly Retrieval from Video Anomaly Detection: New
Benchmarks and Model [70.97446870672069]
Video anomaly detection (VAD) has been paid increasing attention due to its potential applications.
Video Anomaly Retrieval ( VAR) aims to pragmatically retrieve relevant anomalous videos by cross-modalities.
We present two benchmarks, UCFCrime-AR and XD-Violence, constructed on top of prevalent anomaly datasets.
arXiv Detail & Related papers (2023-07-24T06:22:37Z) - Visual anomaly detection in video by variational autoencoder [0.0]
An autoencoder is a neural network that is trained to recreate its input using latent representation of input also called a bottleneck layer.
In this paper we have demonstrated comparison between performance of convolutional LSTM versus a variation convolutional LSTM autoencoder.
arXiv Detail & Related papers (2022-03-08T06:22:04Z) - Anomaly Crossing: A New Method for Video Anomaly Detection as
Cross-domain Few-shot Learning [32.0713939637202]
Video anomaly detection aims to identify abnormal events that occurred in videos.
Most previous approaches learn only from normal videos using unsupervised or semi-supervised methods.
We propose a new learning paradigm by making full use of both normal and abnormal videos for video anomaly detection.
arXiv Detail & Related papers (2021-12-12T20:49:38Z) - Weakly Supervised Video Anomaly Detection via Center-guided
Discriminative Learning [25.787860059872106]
Anomaly detection in surveillance videos is a challenging task due to the diversity of anomalous video content and duration.
We propose an anomaly detection framework, called Anomaly Regression Net (AR-Net), which only requires video-level labels in training stage.
Our method yields a new state-of-the-art result for video anomaly detection on ShanghaiTech dataset.
arXiv Detail & Related papers (2021-04-15T06:41:23Z) - Robust Unsupervised Video Anomaly Detection by Multi-Path Frame
Prediction [61.17654438176999]
We propose a novel and robust unsupervised video anomaly detection method by frame prediction with proper design.
Our proposed method obtains the frame-level AUROC score of 88.3% on the CUHK Avenue dataset.
arXiv Detail & Related papers (2020-11-05T11:34:12Z) - A Self-Reasoning Framework for Anomaly Detection Using Video-Level
Labels [17.615297975503648]
Alous event detection in surveillance videos is a challenging and practical research problem among image and video processing community.
We propose a weakly supervised anomaly detection framework based on deep neural networks which is trained in a self-reasoning fashion using only video-level labels.
The proposed framework has been evaluated on publicly available real-world anomaly detection datasets including UCF-crime, ShanghaiTech and Ped2.
arXiv Detail & Related papers (2020-08-27T02:14:15Z) - Self-trained Deep Ordinal Regression for End-to-End Video Anomaly
Detection [114.9714355807607]
We show that applying self-trained deep ordinal regression to video anomaly detection overcomes two key limitations of existing methods.
We devise an end-to-end trainable video anomaly detection approach that enables joint representation learning and anomaly scoring without manually labeled normal/abnormal data.
arXiv Detail & Related papers (2020-03-15T08:44:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.