Regularity Learning via Explicit Distribution Modeling for Skeletal
Video Anomaly Detection
- URL: http://arxiv.org/abs/2112.03649v2
- Date: Wed, 8 Dec 2021 04:34:47 GMT
- Title: Regularity Learning via Explicit Distribution Modeling for Skeletal
Video Anomaly Detection
- Authors: Shoubin Yu, Zhongyin Zhao, Haoshu Fang, Andong Deng, Haisheng Su,
Dongliang Wang, Weihao Gan, Cewu Lu, Wei Wu
- Abstract summary: A novel Motion Embedder (ME) is proposed to provide a pose motion representation from the probability perspective.
A novel task-specific Spatial-Temporal Transformer (STT) is deployed for self-supervised pose sequence reconstruction.
MoPRL achieves the state-of-the-art performance by an average improvement of 4.7% AUC on several challenging datasets.
- Score: 43.004613173363566
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Anomaly detection in surveillance videos is challenging and important for
ensuring public security. Different from pixel-based anomaly detection methods,
pose-based methods utilize highly-structured skeleton data, which decreases the
computational burden and also avoids the negative impact of background noise.
However, unlike pixel-based methods, which could directly exploit explicit
motion features such as optical flow, pose-based methods suffer from the lack
of alternative dynamic representation. In this paper, a novel Motion Embedder
(ME) is proposed to provide a pose motion representation from the probability
perspective. Furthermore, a novel task-specific Spatial-Temporal Transformer
(STT) is deployed for self-supervised pose sequence reconstruction. These two
modules are then integrated into a unified framework for pose regularity
learning, which is referred to as Motion Prior Regularity Learner (MoPRL).
MoPRL achieves the state-of-the-art performance by an average improvement of
4.7% AUC on several challenging datasets. Extensive experiments validate the
versatility of each proposed module.
Related papers
- Appearance Blur-driven AutoEncoder and Motion-guided Memory Module for Video Anomaly Detection [14.315287192621662]
Video anomaly detection (VAD) often learns the distribution of normal samples and detects the anomaly through measuring significant deviations.
Most VADs cannot cope with cross-dataset validation for new target domains.
We propose a novel VAD method with a motion-guided memory module to achieve cross-dataset validation with zero-shot.
arXiv Detail & Related papers (2024-09-26T07:48:20Z) - Open-Set Deepfake Detection: A Parameter-Efficient Adaptation Method with Forgery Style Mixture [58.60915132222421]
We introduce an approach that is both general and parameter-efficient for face forgery detection.
We design a forgery-style mixture formulation that augments the diversity of forgery source domains.
We show that the designed model achieves state-of-the-art generalizability with significantly reduced trainable parameters.
arXiv Detail & Related papers (2024-08-23T01:53:36Z) - Diffusion-Based Particle-DETR for BEV Perception [94.88305708174796]
Bird-Eye-View (BEV) is one of the most widely-used scene representations for visual perception in Autonomous Vehicles (AVs)
Recent diffusion-based methods offer a promising approach to uncertainty modeling for visual perception but fail to effectively detect small objects in the large coverage of the BEV.
Here, we address this problem by combining the diffusion paradigm with current state-of-the-art 3D object detectors in BEV.
arXiv Detail & Related papers (2023-12-18T09:52:14Z) - Learning from Multi-Perception Features for Real-Word Image
Super-resolution [87.71135803794519]
We propose a novel SR method called MPF-Net that leverages multiple perceptual features of input images.
Our method incorporates a Multi-Perception Feature Extraction (MPFE) module to extract diverse perceptual information.
We also introduce a contrastive regularization term (CR) that improves the model's learning capability.
arXiv Detail & Related papers (2023-05-26T07:35:49Z) - Meta-tuning Loss Functions and Data Augmentation for Few-shot Object
Detection [7.262048441360132]
Few-shot object detection is an emerging topic in the area of few-shot learning and object detection.
We propose a training scheme that allows learning inductive biases that can boost few-shot detection.
The proposed approach yields interpretable loss functions, as opposed to highly parametric and complex few-shot meta-models.
arXiv Detail & Related papers (2023-04-24T15:14:16Z) - Single-photon Image Super-resolution via Self-supervised Learning [6.218646347012887]
Single-Photon Image Super-Resolution (SPISR) aims to recover a high-resolution photon counting cube from a noisy low-resolution one by computational imaging algorithms.
By extending Equivariant Imaging (EI) to single-photon data, we propose a self-supervised learning framework for the SPISR task.
arXiv Detail & Related papers (2023-03-03T15:52:01Z) - Bayesian Nonparametric Submodular Video Partition for Robust Anomaly
Detection [9.145168943972067]
Multiple-instance learning (MIL) provides an effective way to tackle the video anomaly detection problem.
We propose to conduct novel Bayesian non-parametric submodular video partition (BN-SVP) to significantly improve MIL model training.
Our theoretical analysis ensures a strong performance guarantee of the proposed algorithm.
arXiv Detail & Related papers (2022-03-24T04:00:49Z) - Object-centric and memory-guided normality reconstruction for video
anomaly detection [56.64792194894702]
This paper addresses anomaly detection problem for videosurveillance.
Due to the inherent rarity and heterogeneity of abnormal events, the problem is viewed as a normality modeling strategy.
Our model learns object-centric normal patterns without seeing anomalous samples during training.
arXiv Detail & Related papers (2022-03-07T19:28:39Z) - CDN-MEDAL: Two-stage Density and Difference Approximation Framework for
Motion Analysis [3.337126420148156]
We propose a novel, two-stage method of change detection with two convolutional neural networks.
Our two-stage framework contains approximately 3.5K parameters in total but still maintains rapid convergence to intricate motion patterns.
arXiv Detail & Related papers (2021-06-07T16:39:42Z) - Progressive Self-Guided Loss for Salient Object Detection [102.35488902433896]
We present a progressive self-guided loss function to facilitate deep learning-based salient object detection in images.
Our framework takes advantage of adaptively aggregated multi-scale features to locate and detect salient objects effectively.
arXiv Detail & Related papers (2021-01-07T07:33:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.