SDFA: Structure Aware Discriminative Feature Aggregation for Efficient Human Fall Detection in Video
- URL: http://arxiv.org/abs/2503.07008v1
- Date: Mon, 10 Mar 2025 07:46:00 GMT
- Title: SDFA: Structure Aware Discriminative Feature Aggregation for Efficient Human Fall Detection in Video
- Authors: Sania Zahan, Ghulam Mubashar Hassan, Ajmal Mian,
- Abstract summary: Older people are susceptible to fall due to instability in posture and deteriorating health.<n>We propose a fall detection model, coined SDFA, based on human skeletons extracted from low-resolution videos.<n>Our model captures discriminative structural displacements and motion trends using unified joint and motion features projected onto a shared high dimensional space.
- Score: 25.57448683000322
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Older people are susceptible to fall due to instability in posture and deteriorating health. Immediate access to medical support can greatly reduce repercussions. Hence, there is an increasing interest in automated fall detection, often incorporated into a smart healthcare system to provide better monitoring. Existing systems focus on wearable devices which are inconvenient or video monitoring which has privacy concerns. Moreover, these systems provide a limited perspective of their generalization ability as they are tested on datasets containing few activities that have wide disparity in the action space and are easy to differentiate. Complex daily life scenarios pose much greater challenges with activities that overlap in action spaces due to similar posture or motion. To overcome these limitations, we propose a fall detection model, coined SDFA, based on human skeletons extracted from low-resolution videos. The use of skeleton data ensures privacy and low-resolution videos ensures low hardware and computational cost. Our model captures discriminative structural displacements and motion trends using unified joint and motion features projected onto a shared high dimensional space. Particularly, the use of separable convolution combined with a powerful GCN architecture provides improved performance. Extensive experiments on five large-scale datasets with a wide range of evaluation settings show that our model achieves competitive performance with extremely low computational complexity and runs faster than existing models.
Related papers
- UCF-Crime-DVS: A Novel Event-Based Dataset for Video Anomaly Detection with Spiking Neural Networks [7.079697386550486]
Dynamic vision sensors (DVS) capture visual information as discrete events with a very high dynamic range and temporal resolution.
To introduce this rich dynamic information into the surveillance field, we created the first DVS video anomaly detection benchmark, UCF-Crime-DVS.
To fully utilize this new data modality, a multi-scale spiking fusion network (MSF) is designed based on spiking neural networks (SNNs)
arXiv Detail & Related papers (2025-03-17T08:11:26Z) - Modeling Human Skeleton Joint Dynamics for Fall Detection [25.57448683000322]
Fall detection from video streams is not an attractive option for real-life applications due to privacy issues.<n>Key points on the body such as skeleton joints, can convey significant information about motion dynamics.<n>We propose an efficient graph convolution network model that exploitstemporal joint dependencies and dynamics of human skeleton joints for accurate fall detection.
arXiv Detail & Related papers (2025-03-10T05:35:56Z) - Computer-Aided Fall Recognition Using a Three-Stream Spatial-Temporal GCN Model with Adaptive Feature Aggregation [0.5235143203977018]
Prevention of falls is paramount in modern healthcare, particularly for the elderly.
A computer-aided fall detection system is inevitable to save elderly people's lives worldwide.
This paper proposes a novel three-stream spatial-temporal feature-based fall detection system.
arXiv Detail & Related papers (2024-08-22T08:40:04Z) - Towards Evaluating the Robustness of Visual State Space Models [63.14954591606638]
Vision State Space Models (VSSMs) have demonstrated remarkable performance in visual perception tasks.
However, their robustness under natural and adversarial perturbations remains a critical concern.
We present a comprehensive evaluation of VSSMs' robustness under various perturbation scenarios.
arXiv Detail & Related papers (2024-06-13T17:59:44Z) - A Reliable Framework for Human-in-the-Loop Anomaly Detection in Time Series [17.08674819906415]
We introduce HILAD, a novel framework designed to foster a dynamic and bidirectional collaboration between humans and AI.
Through our visual interface, HILAD empowers domain experts to detect, interpret, and correct unexpected model behaviors at scale.
arXiv Detail & Related papers (2024-05-06T07:44:07Z) - XAI-based gait analysis of patients walking with Knee-Ankle-Foot
orthosis using video cameras [1.8749305679160366]
This paper presents a novel system for gait analysis robust to camera movements and providing explanations for its output.
The proposed system employs super-resolution and pose estimation during pre-processing.
It then identifies the seven features - Stride Length, Step Length and Duration of single support of orthotic and non-orthotic leg, Cadence, and Speed.
arXiv Detail & Related papers (2024-02-25T19:05:10Z) - Open-Vocabulary Video Anomaly Detection [57.552523669351636]
Video anomaly detection (VAD) with weak supervision has achieved remarkable performance in utilizing video-level labels to discriminate whether a video frame is normal or abnormal.
Recent studies attempt to tackle a more realistic setting, open-set VAD, which aims to detect unseen anomalies given seen anomalies and normal videos.
This paper takes a step further and explores open-vocabulary video anomaly detection (OVVAD), in which we aim to leverage pre-trained large models to detect and categorize seen and unseen anomalies.
arXiv Detail & Related papers (2023-11-13T02:54:17Z) - An Outlier Exposure Approach to Improve Visual Anomaly Detection
Performance for Mobile Robots [76.36017224414523]
We consider the problem of building visual anomaly detection systems for mobile robots.
Standard anomaly detection models are trained using large datasets composed only of non-anomalous data.
We tackle the problem of exploiting these data to improve the performance of a Real-NVP anomaly detection model.
arXiv Detail & Related papers (2022-09-20T15:18:13Z) - Coherent Loss: A Generic Framework for Stable Video Segmentation [103.78087255807482]
We investigate how a jittering artifact degrades the visual quality of video segmentation results.
We propose a Coherent Loss with a generic framework to enhance the performance of a neural network against jittering artifacts.
arXiv Detail & Related papers (2020-10-25T10:48:28Z) - Low to High Dimensional Modality Hallucination using Aggregated Fields
of View [48.32515709424962]
We argue modality hallucination as one effective way to ensure consistent modality availability.
We present a novel hallucination architecture that aggregates information from multiple fields of view of the local neighborhood.
We also conduct extensive classification and segmentation experiments on UWRGBD and NYUD datasets and demonstrate that hallucination allays the negative effects of the modality loss.
arXiv Detail & Related papers (2020-07-13T03:13:48Z) - SUOD: Accelerating Large-Scale Unsupervised Heterogeneous Outlier
Detection [63.253850875265115]
Outlier detection (OD) is a key machine learning (ML) task for identifying abnormal objects from general samples.
We propose a modular acceleration system, called SUOD, to address it.
arXiv Detail & Related papers (2020-03-11T00:22:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.