Vision-based Fight Detection from Surveillance Cameras
- URL: http://arxiv.org/abs/2002.04355v1
- Date: Tue, 11 Feb 2020 12:56:29 GMT
- Title: Vision-based Fight Detection from Surveillance Cameras
- Authors: \c{S}eymanur Akt{\i}, G\"ozde Ay\c{s}e Tataro\u{g}lu, Haz{\i}m Kemal
Ekenel
- Abstract summary: This paper explores LSTM-based approaches to solve fight scene classification problem.
A new dataset is collected, which consists of fight scenes from surveillance camera videos available at YouTube.
It is observed that the proposed approach, which integrates Xception model, Bi-LSTM, and attention, improves the state-of-the-art accuracy for fight scene classification.
- Score: 6.982738885923204
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Vision-based action recognition is one of the most challenging research
topics of computer vision and pattern recognition. A specific application of
it, namely, detecting fights from surveillance cameras in public areas,
prisons, etc., is desired to quickly get under control these violent incidents.
This paper addresses this research problem and explores LSTM-based approaches
to solve it. Moreover, the attention layer is also utilized. Besides, a new
dataset is collected, which consists of fight scenes from surveillance camera
videos available at YouTube. This dataset is made publicly available. From the
extensive experiments conducted on Hockey Fight, Peliculas, and the newly
collected fight datasets, it is observed that the proposed approach, which
integrates Xception model, Bi-LSTM, and attention, improves the
state-of-the-art accuracy for fight scene classification.
Related papers
- JOSENet: A Joint Stream Embedding Network for Violence Detection in Surveillance Videos [4.94659999696881]
Violence detection in surveillance videos presents additional issues, such as the wide variety of real fight scenes.
We introduce JOSENet, a self-supervised framework that provides outstanding performance for violence detection in surveillance videos.
arXiv Detail & Related papers (2024-05-05T15:01:00Z) - SeaDSC: A video-based unsupervised method for dynamic scene change
detection in unmanned surface vehicles [3.2716252389196288]
This paper outlines our approach to detect dynamic scene changes in Unmanned Surface Vehicles (USVs)
Our objective is to identify significant changes in the dynamic scenes of maritime video data, particularly those scenes that exhibit a high degree of resemblance.
In our system for dynamic scene change detection, we propose completely unsupervised learning method.
arXiv Detail & Related papers (2023-11-20T07:34:01Z) - Survey on video anomaly detection in dynamic scenes with moving cameras [16.849332621082613]
We aim to present the first comprehensive survey on Moving Camera Video Anomaly Detection (MC-VAD)
Our exploration encompasses three application domains: security, urban transportation, and marine environments, which in turn cover six specific tasks.
We compile an extensive list of 25 publicly-available datasets spanning four distinct environments: underwater, water surface, ground, and aerial.
We summarize the types of anomalies these datasets correspond to or contain, and present five main categories of approaches for detecting such anomalies.
arXiv Detail & Related papers (2023-08-14T10:21:06Z) - CCTV-Gun: Benchmarking Handgun Detection in CCTV Images [59.24281591714385]
Gun violence is a critical security problem, and it is imperative for the computer vision community to develop effective gun detection algorithms.
detecting guns in real-world CCTV images remains a challenging and under-explored task.
We present a benchmark, called textbfCCTV-Gun, which addresses the challenges of detecting handguns in real-world CCTV images.
arXiv Detail & Related papers (2023-03-19T16:17:35Z) - Deep Learning for Event-based Vision: A Comprehensive Survey and Benchmarks [55.81577205593956]
Event cameras are bio-inspired sensors that capture the per-pixel intensity changes asynchronously.
Deep learning (DL) has been brought to this emerging field and inspired active research endeavors in mining its potential.
arXiv Detail & Related papers (2023-02-17T14:19:28Z) - Weakly Supervised Two-Stage Training Scheme for Deep Video Fight
Detection Model [0.0]
Fight detection in videos is an emerging deep learning application with today's prevalence of surveillance systems and streaming media.
Previous work has largely relied on action recognition techniques to tackle this problem.
We design the fight detection model as a composition of an action-aware feature extractor and an anomaly score generator.
arXiv Detail & Related papers (2022-09-23T08:29:16Z) - The State of Aerial Surveillance: A Survey [62.198765910573556]
This paper provides a comprehensive overview of human-centric aerial surveillance tasks from a computer vision and pattern recognition perspective.
The main object of interest is humans, where single or multiple subjects are to be detected, identified, tracked, re-identified and have their behavior analyzed.
arXiv Detail & Related papers (2022-01-09T20:13:27Z) - Real Time Action Recognition from Video Footage [0.5219568203653523]
Video surveillance cameras have added a new dimension to detect crime.
This research focuses on integrating state-of-the-art Deep Learning methods to ensure a robust pipeline for autonomous surveillance for detecting violent activities.
arXiv Detail & Related papers (2021-12-13T07:27:41Z) - ASOD60K: Audio-Induced Salient Object Detection in Panoramic Videos [79.05486554647918]
We propose PV-SOD, a new task that aims to segment salient objects from panoramic videos.
In contrast to existing fixation-level or object-level saliency detection tasks, we focus on multi-modal salient object detection (SOD)
We collect the first large-scale dataset, named ASOD60K, which contains 4K-resolution video frames annotated with a six-level hierarchy.
arXiv Detail & Related papers (2021-07-24T15:14:20Z) - Onfocus Detection: Identifying Individual-Camera Eye Contact from
Unconstrained Images [81.64699115587167]
Onfocus detection aims at identifying whether the focus of the individual captured by a camera is on the camera or not.
We build a large-scale onfocus detection dataset, named as the OnFocus Detection In the Wild (OFDIW)
We propose a novel end-to-end deep model, i.e., the eye-context interaction inferring network (ECIIN) for onfocus detection.
arXiv Detail & Related papers (2021-03-29T03:29:09Z) - Multi-Modal Video Forensic Platform for Investigating Post-Terrorist
Attack Scenarios [55.82693757287532]
Large scale Video Analytic Platforms (VAP) assist law enforcement agencies (LEA) in identifying suspects and securing evidence.
We present a video analytic platform that integrates visual and audio analytic modules and fuses information from surveillance cameras and video uploads from eyewitnesses.
arXiv Detail & Related papers (2020-04-02T14:29:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.