Pedestrian Spatio-Temporal Information Fusion For Video Anomaly
Detection
- URL: http://arxiv.org/abs/2211.10052v1
- Date: Fri, 18 Nov 2022 06:41:02 GMT
- Title: Pedestrian Spatio-Temporal Information Fusion For Video Anomaly
Detection
- Authors: Chao Hu, Liqiang Zhu
- Abstract summary: An anomaly detection method is proposed to integrate the information of pedestrians.
Anomaly detection is realized according to the difference between the output frame and the true value.
The experimental results on the CUHK Avenue and ShanghaiTech datasets show that the proposed method is superior to the current mainstream video anomaly detection methods.
- Score: 1.5736899098702974
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Aiming at the problem that the current video anomaly detection cannot fully
use the temporal information and ignore the diversity of normal behavior, an
anomaly detection method is proposed to integrate the spatiotemporal
information of pedestrians. Based on the convolutional autoencoder, the input
frame is compressed and restored through the encoder and decoder. Anomaly
detection is realized according to the difference between the output frame and
the true value. In order to strengthen the characteristic information
connection between continuous video frames, the residual temporal shift module
and the residual channel attention module are introduced to improve the
modeling ability of the network on temporal information and channel
information, respectively. Due to the excessive generalization of convolutional
neural networks, in the memory enhancement modules, the hopping connections of
each codec layer are added to limit autoencoders' ability to represent abnormal
frames too vigorously and improve the anomaly detection accuracy of the
network. In addition, the objective function is modified by a feature
discretization loss, which effectively distinguishes different normal behavior
patterns. The experimental results on the CUHK Avenue and ShanghaiTech datasets
show that the proposed method is superior to the current mainstream video
anomaly detection methods while meeting the real-time requirements.
Related papers
- Abnormal Event Detection In Videos Using Deep Embedding [0.0]
Abnormal event detection or anomaly detection in surveillance videos is currently a challenge because of the diversity of possible events.
We propose an unsupervised approach for video anomaly detection with the aim to jointly optimize the objectives of the deep neural network.
arXiv Detail & Related papers (2024-09-15T17:44:51Z) - Weakly Supervised Video Anomaly Detection and Localization with Spatio-Temporal Prompts [57.01985221057047]
This paper introduces a novel method that learnstemporal prompt embeddings for weakly supervised video anomaly detection and localization (WSVADL) based on pre-trained vision-language models (VLMs)
Our method achieves state-of-theart performance on three public benchmarks for the WSVADL task.
arXiv Detail & Related papers (2024-08-12T03:31:29Z) - Patch Spatio-Temporal Relation Prediction for Video Anomaly Detection [19.643936110623653]
Video Anomaly Detection (VAD) aims to identify abnormalities within a specific context and timeframe.
Recent deep learning-based VAD models have shown promising results by generating high-resolution frames.
We propose a self-supervised learning approach for VAD through an inter-patch relationship prediction task.
arXiv Detail & Related papers (2024-03-28T03:07:16Z) - Dynamic Erasing Network Based on Multi-Scale Temporal Features for
Weakly Supervised Video Anomaly Detection [103.92970668001277]
We propose a Dynamic Erasing Network (DE-Net) for weakly supervised video anomaly detection.
We first propose a multi-scale temporal modeling module, capable of extracting features from segments of varying lengths.
Then, we design a dynamic erasing strategy, which dynamically assesses the completeness of the detected anomalies.
arXiv Detail & Related papers (2023-12-04T09:40:11Z) - A Spatial-Temporal Deformable Attention based Framework for Breast
Lesion Detection in Videos [107.96514633713034]
We propose a spatial-temporal deformable attention based framework, named STNet.
Our STNet introduces a spatial-temporal deformable attention module to perform local spatial-temporal feature fusion.
Experiments on the public breast lesion ultrasound video dataset show that our STNet obtains a state-of-the-art detection performance.
arXiv Detail & Related papers (2023-09-09T07:00:10Z) - A Robust and Explainable Data-Driven Anomaly Detection Approach For
Power Electronics [56.86150790999639]
We present two anomaly detection and classification approaches, namely the Matrix Profile algorithm and anomaly transformer.
The Matrix Profile algorithm is shown to be well suited as a generalizable approach for detecting real-time anomalies in streaming time-series data.
A series of custom filters is created and added to the detector to tune its sensitivity, recall, and detection accuracy.
arXiv Detail & Related papers (2022-09-23T06:09:35Z) - Video Salient Object Detection via Contrastive Features and Attention
Modules [106.33219760012048]
We propose a network with attention modules to learn contrastive features for video salient object detection.
A co-attention formulation is utilized to combine the low-level and high-level features.
We show that the proposed method requires less computation, and performs favorably against the state-of-the-art approaches.
arXiv Detail & Related papers (2021-11-03T17:40:32Z) - A Modular and Unified Framework for Detecting and Localizing Video
Anomalies [30.83924581439373]
We propose a modular and unified approach to the online video anomaly detection and localization problem, called MOVAD.
It consists of a novel transfer learning based plug-and-play architecture, a sequential anomaly detector, a mathematical framework for selecting the detection threshold, and a suitable performance metric for real-time anomalous event detection in videos.
arXiv Detail & Related papers (2021-03-21T04:16:51Z) - Temporal-Channel Transformer for 3D Lidar-Based Video Object Detection
in Autonomous Driving [121.44554957537613]
We propose a new transformer, called Temporal-Channel Transformer, to model the spatial-temporal domain and channel domain relationships for video object detecting from Lidar data.
Specifically, the temporal-channel encoder of the transformer is designed to encode the information of different channels and frames.
We achieve the state-of-the-art performance in grid voxel-based 3D object detection on the nuScenes benchmark.
arXiv Detail & Related papers (2020-11-27T09:35:39Z) - Learning Memory-guided Normality for Anomaly Detection [33.77435699029528]
We present an unsupervised learning approach to anomaly detection that considers the diversity of normal patterns explicitly.
We also present novel feature compactness and separateness losses to train the memory, boosting the discriminative power of both memory items and deeply learned features from normal data.
arXiv Detail & Related papers (2020-03-30T05:30:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.