A Deep Learning Approach to Video Anomaly Detection using Convolutional
Autoencoders
- URL: http://arxiv.org/abs/2311.04351v1
- Date: Tue, 7 Nov 2023 21:23:32 GMT
- Title: A Deep Learning Approach to Video Anomaly Detection using Convolutional
Autoencoders
- Authors: Gopikrishna Pavuluri, Gayathri Annem
- Abstract summary: Our method utilizes a convolutional autoencoder to learn the patterns of normal videos and then compares each frame of a test video to this learned representation.
We evaluated our approach and achieved an overall accuracy of 99.35% on the Ped1 dataset and 97% on the Ped2 dataset.
The results show that our method outperforms other state-of-the-art methods and it can be used in real-world applications for video anomaly detection.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this research we propose a deep learning approach for detecting anomalies
in videos using convolutional autoencoder and decoder neural networks on the
UCSD dataset.Our method utilizes a convolutional autoencoder to learn the
spatiotemporal patterns of normal videos and then compares each frame of a test
video to this learned representation. We evaluated our approach on the UCSD
dataset and achieved an overall accuracy of 99.35% on the Ped1 dataset and
99.77% on the Ped2 dataset, demonstrating the effectiveness of our method for
detecting anomalies in surveillance videos. The results show that our method
outperforms other state-of-the-art methods, and it can be used in real-world
applications for video anomaly detection.
Related papers
- Self-Distilled Masked Auto-Encoders are Efficient Video Anomaly
Detectors [117.61449210940955]
We propose an efficient abnormal event detection model based on a lightweight masked auto-encoder (AE) applied at the video frame level.
We introduce an approach to weight tokens based on motion gradients, thus shifting the focus from the static background scene to the foreground objects.
We generate synthetic abnormal events to augment the training videos, and task the masked AE model to jointly reconstruct the original frames.
arXiv Detail & Related papers (2023-06-21T06:18:05Z) - Mitigating Representation Bias in Action Recognition: Algorithms and
Benchmarks [76.35271072704384]
Deep learning models perform poorly when applied to videos with rare scenes or objects.
We tackle this problem from two different angles: algorithm and dataset.
We show that the debiased representation can generalize better when transferred to other datasets and tasks.
arXiv Detail & Related papers (2022-09-20T00:30:35Z) - Differentiable Frequency-based Disentanglement for Aerial Video Action
Recognition [56.91538445510214]
We present a learning algorithm for human activity recognition in videos.
Our approach is designed for UAV videos, which are mainly acquired from obliquely placed dynamic cameras.
We conduct extensive experiments on the UAV Human dataset and the NEC Drone dataset.
arXiv Detail & Related papers (2022-09-15T22:16:52Z) - Abnormal Event Detection in Urban Surveillance Videos Using GAN and
Transfer Learning [5.092028049119383]
Abnormal event detection (AED) in urban surveillance videos has multiple challenges.
This paper uses generative adversarial networks (GANCSDs) and performs transfer learning algorithms on pre trained convolutional neural network (CNN)
Experimental results show that the proposed method can effectively detect and locate abnormal events in crowd scenes.
arXiv Detail & Related papers (2020-11-19T02:39:35Z) - Robust Unsupervised Video Anomaly Detection by Multi-Path Frame
Prediction [61.17654438176999]
We propose a novel and robust unsupervised video anomaly detection method by frame prediction with proper design.
Our proposed method obtains the frame-level AUROC score of 88.3% on the CUHK Avenue dataset.
arXiv Detail & Related papers (2020-11-05T11:34:12Z) - Video Anomaly Detection Using Pre-Trained Deep Convolutional Neural Nets
and Context Mining [2.0646127669654835]
We show how to use pre-trained convolutional neural net models to perform feature extraction and context mining.
We derive contextual properties from the high-level features to further improve the performance of our video anomaly detection method.
arXiv Detail & Related papers (2020-10-06T00:26:14Z) - Self-Supervised Representation Learning for Visual Anomaly Detection [9.642625267699488]
We consider the problem of anomaly detection in images videos, and present a new visual anomaly detection technique for videos.
We propose a simple self-supervision approach for learning temporal coherence across video frames without the use of any optical flow information.
This intuitive approach shows superior performance of visual anomaly detection compared to numerous methods for images and videos on UCF101 and ILSVRC2015 video datasets.
arXiv Detail & Related papers (2020-06-17T04:37:29Z) - Detecting Forged Facial Videos using convolutional neural network [0.0]
We propose to use smaller (fewer parameters to learn) convolutional neural networks (CNN) for a data-driven approach to forged video detection.
To validate our approach, we investigate the FaceForensics public dataset detailing both frame-based and video-based results.
arXiv Detail & Related papers (2020-05-17T19:04:59Z) - Self-trained Deep Ordinal Regression for End-to-End Video Anomaly
Detection [114.9714355807607]
We show that applying self-trained deep ordinal regression to video anomaly detection overcomes two key limitations of existing methods.
We devise an end-to-end trainable video anomaly detection approach that enables joint representation learning and anomaly scoring without manually labeled normal/abnormal data.
arXiv Detail & Related papers (2020-03-15T08:44:55Z) - Emotions Don't Lie: An Audio-Visual Deepfake Detection Method Using
Affective Cues [75.1731999380562]
We present a learning-based method for detecting real and fake deepfake multimedia content.
We extract and analyze the similarity between the two audio and visual modalities from within the same video.
We compare our approach with several SOTA deepfake detection methods and report per-video AUC of 84.4% on the DFDC and 96.6% on the DF-TIMIT datasets.
arXiv Detail & Related papers (2020-03-14T22:07:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.