GlanceVAD: Exploring Glance Supervision for Label-efficient Video
Anomaly Detection
- URL: http://arxiv.org/abs/2403.06154v2
- Date: Tue, 12 Mar 2024 02:44:36 GMT
- Title: GlanceVAD: Exploring Glance Supervision for Label-efficient Video
Anomaly Detection
- Authors: Huaxin Zhang, Xiang Wang, Xiaohao Xu, Xiaonan Huang, Chuchu Han,
Yuehuan Wang, Changxin Gao, Shanjun Zhang, Nong Sang
- Abstract summary: We present a novel labeling paradigm, termed "glance annotation", to achieve a better balance between anomaly detection accuracy and annotation cost.
To assess its effectiveness, we manually annotate the glance annotations for two standard video anomaly detection datasets: UCF-Crime and XD-Violence.
We also propose a customized GlanceVAD method, that leverages ssian kernels as the basic unit to compose the temporal anomaly distribution.
- Score: 36.73022075416943
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In recent years, video anomaly detection has been extensively investigated in
both unsupervised and weakly supervised settings to alleviate costly temporal
labeling. Despite significant progress, these methods still suffer from
unsatisfactory results such as numerous false alarms, primarily due to the
absence of precise temporal anomaly annotation. In this paper, we present a
novel labeling paradigm, termed "glance annotation", to achieve a better
balance between anomaly detection accuracy and annotation cost. Specifically,
glance annotation is a random frame within each abnormal event, which can be
easily accessed and is cost-effective. To assess its effectiveness, we manually
annotate the glance annotations for two standard video anomaly detection
datasets: UCF-Crime and XD-Violence. Additionally, we propose a customized
GlanceVAD method, that leverages gaussian kernels as the basic unit to compose
the temporal anomaly distribution, enabling the learning of diverse and robust
anomaly representations from the glance annotations. Through comprehensive
analysis and experiments, we verify that the proposed labeling paradigm can
achieve an excellent trade-off between annotation cost and model performance.
Extensive experimental results also demonstrate the effectiveness of our
GlanceVAD approach, which significantly outperforms existing advanced
unsupervised and weakly supervised methods. Code and annotations will be
publicly available at https://github.com/pipixin321/GlanceVAD.
Related papers
- BatchNorm-based Weakly Supervised Video Anomaly Detection [117.11382325721016]
In weakly supervised video anomaly detection, temporal features of abnormal events often exhibit outlier characteristics.
We propose a novel method, BN-WVAD, which incorporates BatchNorm into WVAD.
The proposed BN-WVAD model demonstrates state-of-the-art performance on UCF-Crime with an AUC of 87.24%, and XD-Violence, where AP reaches up to 84.93%.
arXiv Detail & Related papers (2023-11-26T17:47:57Z) - Few-shot Message-Enhanced Contrastive Learning for Graph Anomaly
Detection [15.757864894708364]
Graph anomaly detection plays a crucial role in identifying exceptional instances in graph data that deviate significantly from the majority.
We propose a novel few-shot Graph Anomaly Detection model called FMGAD.
We show that FMGAD can achieve better performance than other state-of-the-art methods, regardless of artificially injected anomalies or domain-organic anomalies.
arXiv Detail & Related papers (2023-11-17T07:49:20Z) - Open-Vocabulary Video Anomaly Detection [57.552523669351636]
Video anomaly detection (VAD) with weak supervision has achieved remarkable performance in utilizing video-level labels to discriminate whether a video frame is normal or abnormal.
Recent studies attempt to tackle a more realistic setting, open-set VAD, which aims to detect unseen anomalies given seen anomalies and normal videos.
This paper takes a step further and explores open-vocabulary video anomaly detection (OVVAD), in which we aim to leverage pre-trained large models to detect and categorize seen and unseen anomalies.
arXiv Detail & Related papers (2023-11-13T02:54:17Z) - CL-Flow:Strengthening the Normalizing Flows by Contrastive Learning for
Better Anomaly Detection [1.951082473090397]
We propose a self-supervised anomaly detection approach that combines contrastive learning with 2D-Flow.
Compared to mainstream unsupervised approaches, our self-supervised method demonstrates superior detection accuracy, fewer additional model parameters, and faster inference speed.
Our approach showcases new state-of-the-art results, achieving a performance of 99.6% in image-level AUROC on the MVTecAD dataset and 96.8% in image-level AUROC on the BTAD dataset.
arXiv Detail & Related papers (2023-11-12T10:07:03Z) - A Coarse-to-Fine Pseudo-Labeling (C2FPL) Framework for Unsupervised
Video Anomaly Detection [4.494911384096143]
Detection of anomalous events in videos is an important problem in applications such as surveillance.
We propose a simple-but-effective two-stage pseudo-label generation framework that produces segment-level (normal/anomaly) pseudo-labels.
The proposed coarse-to-fine pseudo-label generator employs carefully-designed hierarchical divisive clustering and statistical hypothesis testing.
arXiv Detail & Related papers (2023-10-26T17:59:19Z) - CARLA: Self-supervised Contrastive Representation Learning for Time Series Anomaly Detection [53.83593870825628]
One main challenge in time series anomaly detection (TSAD) is the lack of labelled data in many real-life scenarios.
Most of the existing anomaly detection methods focus on learning the normal behaviour of unlabelled time series in an unsupervised manner.
We introduce a novel end-to-end self-supervised ContrAstive Representation Learning approach for time series anomaly detection.
arXiv Detail & Related papers (2023-08-18T04:45:56Z) - MGFN: Magnitude-Contrastive Glance-and-Focus Network for
Weakly-Supervised Video Anomaly Detection [39.923871347007875]
We propose a novel glance and focus network to integrate spatial-temporal information for accurate anomaly detection.
Existing approaches that use feature magnitudes to represent the degree of anomalies typically ignore the effects of scene variations.
We propose the Feature Amplification Mechanism and a Magnitude Contrastive Loss to enhance the discriminativeness of feature magnitudes for detecting anomalies.
arXiv Detail & Related papers (2022-11-28T07:10:36Z) - A Video Anomaly Detection Framework based on Appearance-Motion Semantics
Representation Consistency [18.06814233420315]
We propose a framework that uses normal data's appearance and motion semantic representation consistency to handle anomaly detection.
We design a two-stream encoder to encode the appearance and motion information representations of normal samples.
Lower consistency of appearance and motion features of anomalous samples can be used to generate predicted frames with larger reconstruction error.
arXiv Detail & Related papers (2022-04-08T15:59:57Z) - WSSOD: A New Pipeline for Weakly- and Semi-Supervised Object Detection [75.80075054706079]
We propose a weakly- and semi-supervised object detection framework (WSSOD)
An agent detector is first trained on a joint dataset and then used to predict pseudo bounding boxes on weakly-annotated images.
The proposed framework demonstrates remarkable performance on PASCAL-VOC and MSCOCO benchmark, achieving a high performance comparable to those obtained in fully-supervised settings.
arXiv Detail & Related papers (2021-05-21T11:58:50Z) - Robust Unsupervised Video Anomaly Detection by Multi-Path Frame
Prediction [61.17654438176999]
We propose a novel and robust unsupervised video anomaly detection method by frame prediction with proper design.
Our proposed method obtains the frame-level AUROC score of 88.3% on the CUHK Avenue dataset.
arXiv Detail & Related papers (2020-11-05T11:34:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.