Pose2Gait: Extracting Gait Features from Monocular Video of Individuals
with Dementia
- URL: http://arxiv.org/abs/2308.11484v1
- Date: Tue, 22 Aug 2023 14:59:17 GMT
- Title: Pose2Gait: Extracting Gait Features from Monocular Video of Individuals
with Dementia
- Authors: Caroline Malin-Mayor, Vida Adeli, Andrea Sabo, Sergey Noritsyn,
Carolina Gorodetsky, Alfonso Fasano, Andrea Iaboni, Babak Taati
- Abstract summary: Video-based ambient monitoring of gait for older adults with dementia has the potential to detect negative changes in health.
Computer vision-based pose tracking models can process video data automatically and extract joint locations.
These models are not optimized for gait analysis on older adults or clinical populations.
- Score: 3.2739089842471136
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Video-based ambient monitoring of gait for older adults with dementia has the
potential to detect negative changes in health and allow clinicians and
caregivers to intervene early to prevent falls or hospitalizations. Computer
vision-based pose tracking models can process video data automatically and
extract joint locations; however, publicly available models are not optimized
for gait analysis on older adults or clinical populations. In this work we
train a deep neural network to map from a two dimensional pose sequence,
extracted from a video of an individual walking down a hallway toward a
wall-mounted camera, to a set of three-dimensional spatiotemporal gait features
averaged over the walking sequence. The data of individuals with dementia used
in this work was captured at two sites using a wall-mounted system to collect
the video and depth information used to train and evaluate our model. Our
Pose2Gait model is able to extract velocity and step length values from the
video that are correlated with the features from the depth camera, with
Spearman's correlation coefficients of .83 and .60 respectively, showing that
three dimensional spatiotemporal features can be predicted from monocular
video. Future work remains to improve the accuracy of other features, such as
step time and step width, and test the utility of the predicted values for
detecting meaningful changes in gait during longitudinal ambient monitoring.
Related papers
- Occlusion-Aware 3D Motion Interpretation for Abnormal Behavior Detection [10.782354892545651]
We present OAD2D, which discriminates against motion abnormalities based on reconstructing 3D coordinates of mesh vertices and human joints from monocular videos.
We reformulate the abnormal posture estimation by coupling it with Motion to Text (M2T) model in which, the VQVAE is employed to quantize motion features.
Our approach demonstrates the robustness of abnormal behavior detection against severe and self-occlusions, as it reconstructs human motion trajectories in global coordinates.
arXiv Detail & Related papers (2024-07-23T18:41:16Z) - Coordinate Transformer: Achieving Single-stage Multi-person Mesh
Recovery from Videos [91.44553585470688]
Multi-person 3D mesh recovery from videos is a critical first step towards automatic perception of group behavior in virtual reality, physical therapy and beyond.
We propose the Coordinate transFormer (CoordFormer) that directly models multi-person spatial-temporal relations and simultaneously performs multi-mesh recovery in an end-to-end manner.
Experiments on the 3DPW dataset demonstrate that CoordFormer significantly improves the state-of-the-art, outperforming the previously best results by 4.2%, 8.8% and 4.7% according to the MPJPE, PAMPJPE, and PVE metrics, respectively.
arXiv Detail & Related papers (2023-08-20T18:23:07Z) - Kinematic-aware Hierarchical Attention Network for Human Pose Estimation
in Videos [17.831839654593452]
Previous-based human pose estimation methods have shown promising results by leveraging features of consecutive frames.
Most approaches compromise accuracy to jitter and do not comprehend the temporal aspects of human motion.
We design an architecture that exploits kinematic keypoint features.
arXiv Detail & Related papers (2022-11-29T01:46:11Z) - MixSTE: Seq2seq Mixed Spatio-Temporal Encoder for 3D Human Pose
Estimation in Video [75.23812405203778]
Recent solutions have been introduced to estimate 3D human pose from 2D keypoint sequence by considering body joints among all frames globally to learn-temporal correlation.
We propose Mix Mix, which has temporal transformer block to separately model the temporal motion of each joint and a transformer block inter-joint spatial correlation.
In addition, the network output is extended from the central frame to entire frames of input video, improving the coherence between the input and output benchmarks.
arXiv Detail & Related papers (2022-03-02T04:20:59Z) - Real-Time Human Pose Estimation on a Smart Walker using Convolutional
Neural Networks [4.076099054649463]
We present a novel approach to patient monitoring and data-driven human-in-the-loop control in the context of smart walkers.
It is able to extract a complete and compact body representation in real-time and from inexpensive sensors.
Despite promising results, more data should be collected on users with impairments to assess its performance as a rehabilitation tool in real-world scenarios.
arXiv Detail & Related papers (2021-06-28T14:11:48Z) - Spatial-Temporal Correlation and Topology Learning for Person
Re-Identification in Videos [78.45050529204701]
We propose a novel framework to pursue discriminative and robust representation by modeling cross-scale spatial-temporal correlation.
CTL utilizes a CNN backbone and a key-points estimator to extract semantic local features from human body.
It explores a context-reinforced topology to construct multi-scale graphs by considering both global contextual information and physical connections of human body.
arXiv Detail & Related papers (2021-04-15T14:32:12Z) - Self-Attentive 3D Human Pose and Shape Estimation from Videos [82.63503361008607]
We present a video-based learning algorithm for 3D human pose and shape estimation.
We exploit temporal information in videos and propose a self-attention module.
We evaluate our method on the 3DPW, MPI-INF-3DHP, and Human3.6M datasets.
arXiv Detail & Related papers (2021-03-26T00:02:19Z) - Towards Real-time Drowsiness Detection for Elderly Care [0.0]
This paper produces a proof of concept extracting drowsiness information from videos to help elderly living on their own.
To yawning, eyelid and head movement over time, we extracted 3000 images from videos for training and testing of deep learning models integrated with OpenCV.
arXiv Detail & Related papers (2020-10-21T05:48:59Z) - A Graph Attention Spatio-temporal Convolutional Network for 3D Human
Pose Estimation in Video [7.647599484103065]
We improve the learning of constraints in human skeleton by modeling local global spatial information via attention mechanisms.
Our approach effectively mitigates depth ambiguity and self-occlusion, generalizes to half upper body estimation, and achieves competitive performance on 2D-to-3D video pose estimation.
arXiv Detail & Related papers (2020-03-11T14:54:40Z) - Road Curb Detection and Localization with Monocular Forward-view Vehicle
Camera [74.45649274085447]
We propose a robust method for estimating road curb 3D parameters using a calibrated monocular camera equipped with a fisheye lens.
Our approach is able to estimate the vehicle to curb distance in real time with mean accuracy of more than 90%.
arXiv Detail & Related papers (2020-02-28T00:24:18Z) - Anatomy-aware 3D Human Pose Estimation with Bone-based Pose
Decomposition [92.99291528676021]
Instead of directly regressing the 3D joint locations, we decompose the task into bone direction prediction and bone length prediction.
Our motivation is the fact that the bone lengths of a human skeleton remain consistent across time.
Our full model outperforms the previous best results on Human3.6M and MPI-INF-3DHP datasets.
arXiv Detail & Related papers (2020-02-24T15:49:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.