Grid HTM: Hierarchical Temporal Memory for Anomaly Detection in Videos
- URL: http://arxiv.org/abs/2205.15407v1
- Date: Mon, 30 May 2022 20:10:23 GMT
- Title: Grid HTM: Hierarchical Temporal Memory for Anomaly Detection in Videos
- Authors: Vladimir Monakhov, Vajira Thambawita, P{\aa}l Halvorsen, Michael A.
Riegler
- Abstract summary: The interest for video anomaly detection systems has gained traction for the past few years.
Current approaches use deep learning to perform anomaly detection in videos, but this approach has multiple problems.
We introduce a novel version of HTM, namely, Grid HTM, which is an HTM-based architecture specifically for anomaly detection in complex videos.
- Score: 1.0013553984400492
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The interest for video anomaly detection systems has gained traction for the
past few years. The current approaches use deep learning to perform anomaly
detection in videos, but this approach has multiple problems. For starters,
deep learning in general has issues with noise, concept drift, explainability,
and training data volumes. Additionally, anomaly detection in itself is a
complex task and faces challenges such as unknowness, heterogeneity, and class
imbalance. Anomaly detection using deep learning is therefore mainly
constrained to generative models such as generative adversarial networks and
autoencoders due to their unsupervised nature, but even they suffer from
general deep learning issues and are hard to train properly. In this paper, we
explore the capabilities of the Hierarchical Temporal Memory (HTM) algorithm to
perform anomaly detection in videos, as it has favorable properties such as
noise tolerance and online learning which combats concept drift. We introduce a
novel version of HTM, namely, Grid HTM, which is an HTM-based architecture
specifically for anomaly detection in complex videos such as surveillance
footage.
Related papers
- VANE-Bench: Video Anomaly Evaluation Benchmark for Conversational LMMs [64.60035916955837]
VANE-Bench is a benchmark designed to assess the proficiency of Video-LMMs in detecting anomalies and inconsistencies in videos.
Our dataset comprises an array of videos synthetically generated using existing state-of-the-art text-to-video generation models.
We evaluate nine existing Video-LMMs, both open and closed sources, on this benchmarking task and find that most of the models encounter difficulties in effectively identifying the subtle anomalies.
arXiv Detail & Related papers (2024-06-14T17:59:01Z) - Dynamic Erasing Network Based on Multi-Scale Temporal Features for
Weakly Supervised Video Anomaly Detection [103.92970668001277]
We propose a Dynamic Erasing Network (DE-Net) for weakly supervised video anomaly detection.
We first propose a multi-scale temporal modeling module, capable of extracting features from segments of varying lengths.
Then, we design a dynamic erasing strategy, which dynamically assesses the completeness of the detected anomalies.
arXiv Detail & Related papers (2023-12-04T09:40:11Z) - Video Anomaly Detection using GAN [0.0]
This thesis study aims to offer the solution for this use case so that human resources won't be required to keep an eye out for any unusual activity in the surveillance system records.
We have developed a novel generative adversarial network (GAN) based anomaly detection model.
arXiv Detail & Related papers (2023-11-23T16:41:30Z) - Open-Vocabulary Video Anomaly Detection [57.552523669351636]
Video anomaly detection (VAD) with weak supervision has achieved remarkable performance in utilizing video-level labels to discriminate whether a video frame is normal or abnormal.
Recent studies attempt to tackle a more realistic setting, open-set VAD, which aims to detect unseen anomalies given seen anomalies and normal videos.
This paper takes a step further and explores open-vocabulary video anomaly detection (OVVAD), in which we aim to leverage pre-trained large models to detect and categorize seen and unseen anomalies.
arXiv Detail & Related papers (2023-11-13T02:54:17Z) - Self-Supervised Masked Convolutional Transformer Block for Anomaly
Detection [122.4894940892536]
We present a novel self-supervised masked convolutional transformer block (SSMCTB) that comprises the reconstruction-based functionality at a core architectural level.
In this work, we extend our previous self-supervised predictive convolutional attentive block (SSPCAB) with a 3D masked convolutional layer, a transformer for channel-wise attention, as well as a novel self-supervised objective based on Huber loss.
arXiv Detail & Related papers (2022-09-25T04:56:10Z) - Deep Video Anomaly Detection: Opportunities and Challenges [12.077052764803161]
Anomaly detection is a popular and vital task in various research contexts.
Deep learning has shown its capacity in a number of domains, ranging from acoustics, images, to natural language processing.
There are numerous advantages if such intelligent systems could be realised in our daily lives.
arXiv Detail & Related papers (2021-10-11T08:41:51Z) - Weakly Supervised Video Anomaly Detection via Center-guided
Discriminative Learning [25.787860059872106]
Anomaly detection in surveillance videos is a challenging task due to the diversity of anomalous video content and duration.
We propose an anomaly detection framework, called Anomaly Regression Net (AR-Net), which only requires video-level labels in training stage.
Our method yields a new state-of-the-art result for video anomaly detection on ShanghaiTech dataset.
arXiv Detail & Related papers (2021-04-15T06:41:23Z) - Anomaly Detection in Video via Self-Supervised and Multi-Task Learning [113.81927544121625]
Anomaly detection in video is a challenging computer vision problem.
In this paper, we approach anomalous event detection in video through self-supervised and multi-task learning at the object level.
arXiv Detail & Related papers (2020-11-15T10:21:28Z) - Robust Unsupervised Video Anomaly Detection by Multi-Path Frame
Prediction [61.17654438176999]
We propose a novel and robust unsupervised video anomaly detection method by frame prediction with proper design.
Our proposed method obtains the frame-level AUROC score of 88.3% on the CUHK Avenue dataset.
arXiv Detail & Related papers (2020-11-05T11:34:12Z) - Video Anomaly Detection Using Pre-Trained Deep Convolutional Neural Nets
and Context Mining [2.0646127669654835]
We show how to use pre-trained convolutional neural net models to perform feature extraction and context mining.
We derive contextual properties from the high-level features to further improve the performance of our video anomaly detection method.
arXiv Detail & Related papers (2020-10-06T00:26:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.