Detecting Violence in Video Based on Deep Features Fusion Technique
- URL: http://arxiv.org/abs/2204.07443v1
- Date: Fri, 15 Apr 2022 12:51:20 GMT
- Title: Detecting Violence in Video Based on Deep Features Fusion Technique
- Authors: Heyam M. Bin Jahlan and Lamiaa A. Elrefaei
- Abstract summary: This work proposed a novel method to detect violence using a fusion tech-nique of two convolutional neural networks (CNNs)
The performance of the proposed method is evaluated using three standard benchmark datasets in terms of detection accuracy.
- Score: 0.30458514384586394
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: With the rapid growth of surveillance cameras in many public places to
mon-itor human activities such as in malls, streets, schools and, prisons,
there is a strong demand for such systems to detect violence events
automatically. Au-tomatic analysis of video to detect violence is significant
for law enforce-ment. Moreover, it helps to avoid any social, economic and
environmental damages. Mostly, all systems today require manual human
supervisors to de-tect violence scenes in the video which is inefficient and
inaccurate. in this work, we interest in physical violence that involved two
persons or more. This work proposed a novel method to detect violence using a
fusion tech-nique of two significantly different convolutional neural networks
(CNNs) which are AlexNet and SqueezeNet networks. Each network followed by
separate Convolution Long Short Term memory (ConvLSTM) to extract ro-bust and
richer features from a video in the final hidden state. Then, making a fusion
of these two obtained states and fed to the max-pooling layer. Final-ly,
features were classified using a series of fully connected layers and soft-max
classifier. The performance of the proposed method is evaluated using three
standard benchmark datasets in terms of detection accuracy: Hockey Fight
dataset, Movie dataset and Violent Flow dataset. The results show an accuracy
of 97%, 100%, and 96% respectively. A comparison of the results with the state
of the art techniques revealed the promising capability of the proposed method
in recognizing violent videos.
Related papers
- Optimizing Violence Detection in Video Classification Accuracy through 3D Convolutional Neural Networks [0.0]
This study is to identify how many frames should be analyzed at a time in order to optimize a violence detection model's accuracy.
Previous violence classification models have been created, but their application to live footage may be flawed.
The greatest validation accuracy was 94.87% and occurred with the model that analyzed three frames at a time.
arXiv Detail & Related papers (2024-11-02T19:29:01Z) - Violence detection in videos using deep recurrent and convolutional neural networks [0.0]
We propose a deep learning architecture for violence detection which combines both recurrent neural networks (RNNs) and 2-dimensional convolutional neural networks (2D CNN)
In addition to video frames, we use optical flow computed using the captured sequences.
The proposed approaches reach the same level as the state-of-the-art techniques and sometime surpass them.
arXiv Detail & Related papers (2024-09-11T19:21:51Z) - CCTV-Gun: Benchmarking Handgun Detection in CCTV Images [59.24281591714385]
Gun violence is a critical security problem, and it is imperative for the computer vision community to develop effective gun detection algorithms.
detecting guns in real-world CCTV images remains a challenging and under-explored task.
We present a benchmark, called textbfCCTV-Gun, which addresses the challenges of detecting handguns in real-world CCTV images.
arXiv Detail & Related papers (2023-03-19T16:17:35Z) - TempNet: Temporal Attention Towards the Detection of Animal Behaviour in
Videos [63.85815474157357]
We propose an efficient computer vision- and deep learning-based method for the detection of biological behaviours in videos.
TempNet uses an encoder bridge and residual blocks to maintain model performance with a two-staged, spatial, then temporal, encoder.
We demonstrate its application to the detection of sablefish (Anoplopoma fimbria) startle events.
arXiv Detail & Related papers (2022-11-17T23:55:12Z) - Two-stream Multi-dimensional Convolutional Network for Real-time
Violence Detection [0.0]
This work presents a novel architecture for violence detection called Two-stream Multi-dimensional Convolutional Network (2s-MDCN)
Our proposed method extracts temporal and spatial information independently by 1D, 2D, and 3D convolutions.
Our models obtained state-of-the-art accuracy of 89.7% on the largest violence detection benchmark dataset.
arXiv Detail & Related papers (2022-11-08T14:04:47Z) - Video Manipulations Beyond Faces: A Dataset with Human-Machine Analysis [60.13902294276283]
We present VideoSham, a dataset consisting of 826 videos (413 real and 413 manipulated).
Many of the existing deepfake datasets focus exclusively on two types of facial manipulations -- swapping with a different subject's face or altering the existing face.
Our analysis shows that state-of-the-art manipulation detection algorithms only work for a few specific attacks and do not scale well on VideoSham.
arXiv Detail & Related papers (2022-07-26T17:39:04Z) - SSIVD-Net: A Novel Salient Super Image Classification & Detection
Technique for Weaponized Violence [3.651114792588495]
Detection of violence and weaponized violence in CCTV footage requires a comprehensive approach.
We introduce the emphSmart-City CCTV Violence Detection (SCVD) dataset.
We propose a novel technique called emphSSIVD-Net (textbfSalient-textbfSuper-textbfImage for textbfViolence textbfDetection)
arXiv Detail & Related papers (2022-07-26T12:31:01Z) - Real Time Action Recognition from Video Footage [0.5219568203653523]
Video surveillance cameras have added a new dimension to detect crime.
This research focuses on integrating state-of-the-art Deep Learning methods to ensure a robust pipeline for autonomous surveillance for detecting violent activities.
arXiv Detail & Related papers (2021-12-13T07:27:41Z) - Adversarially robust deepfake media detection using fused convolutional
neural network predictions [79.00202519223662]
Current deepfake detection systems struggle against unseen data.
We employ three different deep Convolutional Neural Network (CNN) models to classify fake and real images extracted from videos.
The proposed technique outperforms state-of-the-art models with 96.5% accuracy.
arXiv Detail & Related papers (2021-02-11T11:28:00Z) - Robust Unsupervised Video Anomaly Detection by Multi-Path Frame
Prediction [61.17654438176999]
We propose a novel and robust unsupervised video anomaly detection method by frame prediction with proper design.
Our proposed method obtains the frame-level AUROC score of 88.3% on the CUHK Avenue dataset.
arXiv Detail & Related papers (2020-11-05T11:34:12Z) - VideoForensicsHQ: Detecting High-quality Manipulated Face Videos [77.60295082172098]
We show how the performance of forgery detectors depends on the presence of artefacts that the human eye can see.
We introduce a new benchmark dataset for face video forgery detection, of unprecedented quality.
arXiv Detail & Related papers (2020-05-20T21:17:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.