OmniFall: A Unified Staged-to-Wild Benchmark for Human Fall Detection
- URL: http://arxiv.org/abs/2505.19889v1
- Date: Mon, 26 May 2025 12:19:11 GMT
- Title: OmniFall: A Unified Staged-to-Wild Benchmark for Human Fall Detection
- Authors: David Schneider, Zdravko Marinov, Rafael Baur, Zeyun Zhong, Rodi Düger, Rainer Stiefelhagen,
- Abstract summary: We introduce OmniFall, unifying eight public fall detection datasets.<n>For real-world evaluation we curate OOPS-Fall from genuine accident videos.<n>Experiments with frozen pre-trained backbones such as I3D or VideoMAE reveal significant performance gaps between in-distribution and in-the-wild scenarios.
- Score: 23.863250825189795
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Current video-based fall detection research mostly relies on small, staged datasets with significant domain biases concerning background, lighting, and camera setup resulting in unknown real-world performance. We introduce OmniFall, unifying eight public fall detection datasets (roughly 14 h of recordings, roughly 42 h of multiview data, 101 subjects, 29 camera views) under a consistent ten-class taxonomy with standardized evaluation protocols. Our benchmark provides complete video segmentation labels and enables fair cross-dataset comparison previously impossible with incompatible annotation schemes. For real-world evaluation we curate OOPS-Fall from genuine accident videos and establish a staged-to-wild protocol measuring generalization from controlled to uncontrolled environments. Experiments with frozen pre-trained backbones such as I3D or VideoMAE reveal significant performance gaps between in-distribution and in-the-wild scenarios, highlighting critical challenges in developing robust fall detection systems. OmniFall Dataset at https://huggingface.co/datasets/simplexsigil2/omnifall , Code at https://github.com/simplexsigil/omnifall-experiments
Related papers
- Track Any Anomalous Object: A Granular Video Anomaly Detection Pipeline [63.96226274616927]
A new framework called Track Any Anomalous Object (TAO) introduces a granular video anomaly detection pipeline.<n>Unlike methods that assign anomaly scores to every pixel, our approach transforms the problem into pixel-level tracking of anomalous objects.<n>Experiments demonstrate that TAO sets new benchmarks in accuracy and robustness.
arXiv Detail & Related papers (2025-06-05T15:49:39Z) - Towards Scalable Video Anomaly Retrieval: A Synthetic Video-Text Benchmark [26.948237287675116]
Video anomaly retrieval aims to localize anomalous events in videos using natural language queries to facilitate public safety.<n>Existing datasets suffer from data scarcity due to the long-tail nature of real-world anomalies, and privacy constraints that impede large-scale collection.<n>We introduce SVTA (Synthetic Video-Text Anomaly benchmark), the first large-scale dataset for cross-modal anomaly retrieval.
arXiv Detail & Related papers (2025-06-02T09:23:58Z) - SOTA: Spike-Navigated Optimal TrAnsport Saliency Region Detection in Composite-bias Videos [50.51658520045165]
Spike-d Optimal TrAnsport Saliency Region Detection (SOTA) is a framework that leverages the strengths of spike cameras while mitigating biases in both spatial and temporal dimensions.<n>Our method introduces Spike-based Micro-debias (SM) to capture subtle frame-to-frame variations.<n>SOTA refines predictions by reducing inconsistencies across diverse conditions.
arXiv Detail & Related papers (2025-05-01T08:30:40Z) - VANE-Bench: Video Anomaly Evaluation Benchmark for Conversational LMMs [64.60035916955837]
VANE-Bench is a benchmark designed to assess the proficiency of Video-LMMs in detecting anomalies and inconsistencies in videos.<n>Our dataset comprises an array of videos synthetically generated using existing state-of-the-art text-to-video generation models.<n>We evaluate nine existing Video-LMMs, both open and closed sources, on this benchmarking task and find that most of the models encounter difficulties in effectively identifying the subtle anomalies.
arXiv Detail & Related papers (2024-06-14T17:59:01Z) - Cutup and Detect: Human Fall Detection on Cutup Untrimmed Videos Using a
Large Foundational Video Understanding Model [0.0]
This work explores the performance of a large video understanding foundation model on the downstream task of human fall detection on untrimmed video.
A method for temporal action localization that relies on a simple cutup of untrimmed videos is demonstrated.
The results are promising for real-time application, and the falls are detected on video level with a state-of-the-art 0.96 F1 score on the HQFSD dataset.
arXiv Detail & Related papers (2024-01-29T16:37:00Z) - EGOFALLS: A visual-audio dataset and benchmark for fall detection using
egocentric cameras [0.16317061277456998]
Falls are significant and often fatal for vulnerable populations such as the elderly.
Previous works have addressed the detection of falls by relying on data capture by a single sensor, images or accelerometers.
In this work, we rely on multimodal descriptors extracted from videos captured by egocentric cameras.
arXiv Detail & Related papers (2023-09-08T20:14:25Z) - CHAD: Charlotte Anomaly Dataset [2.6774008509840996]
We present the Charlotte Anomaly dataset (CHAD) for video anomaly detection.
CHAD is the first anomaly dataset to include bounding box, identity, and pose annotations for each actor.
With four camera views and over 1.15 million frames, CHAD is the largest fully annotated anomaly detection dataset.
arXiv Detail & Related papers (2022-12-19T06:05:34Z) - Shuffled Patch-Wise Supervision for Presentation Attack Detection [12.031796234206135]
Face anti-spoofing is essential to prevent false facial verification by using a photo, video, mask, or a different substitute for an authorized person's face.
Most presentation attack detection systems suffer from overfitting, where they achieve near-perfect scores on a single dataset but fail on a different dataset with more realistic data.
We propose a new PAD approach, which combines pixel-wise binary supervision with patch-based CNN.
arXiv Detail & Related papers (2021-09-08T08:14:13Z) - Robust Unsupervised Video Anomaly Detection by Multi-Path Frame
Prediction [61.17654438176999]
We propose a novel and robust unsupervised video anomaly detection method by frame prediction with proper design.
Our proposed method obtains the frame-level AUROC score of 88.3% on the CUHK Avenue dataset.
arXiv Detail & Related papers (2020-11-05T11:34:12Z) - A Background-Agnostic Framework with Adversarial Training for Abnormal
Event Detection in Video [120.18562044084678]
Abnormal event detection in video is a complex computer vision problem that has attracted significant attention in recent years.
We propose a background-agnostic framework that learns from training videos containing only normal events.
arXiv Detail & Related papers (2020-08-27T18:39:24Z) - Uncertainty-Aware Weakly Supervised Action Detection from Untrimmed
Videos [82.02074241700728]
In this paper, we present a prohibitive-level action recognition model that is trained with only video-frame labels.
Our method per person detectors have been trained on large image datasets within Multiple Instance Learning framework.
We show how we can apply our method in cases where the standard Multiple Instance Learning assumption, that each bag contains at least one instance with the specified label, is invalid.
arXiv Detail & Related papers (2020-07-21T10:45:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.