A Spatio-temporal Attention-based Model for Infant Movement Assessment
from Videos
- URL: http://arxiv.org/abs/2105.09783v1
- Date: Thu, 20 May 2021 14:31:54 GMT
- Title: A Spatio-temporal Attention-based Model for Infant Movement Assessment
from Videos
- Authors: Binh Nguyen-Thai, Vuong Le, Catherine Morgan, Nadia Badawi, Truyen
Tran, and Svetha Venkatesh
- Abstract summary: We develop a new method for fidgety movement assessment using human poses extracted from short clips.
Human poses capture only relevant motion profiles of joints and limbs and are free from irrelevant appearance artifacts.
Our experiments show that the proposed method achieves the ROC-AUC score of 81.87%, significantly outperforming existing competing methods with better interpretability.
- Score: 44.71923220732036
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The absence or abnormality of fidgety movements of joints or limbs is
strongly indicative of cerebral palsy in infants. Developing computer-based
methods for assessing infant movements in videos is pivotal for improved
cerebral palsy screening. Most existing methods use appearance-based features
and are thus sensitive to strong but irrelevant signals caused by background
clutter or a moving camera. Moreover, these features are computed over the
whole frame, thus they measure gross whole body movements rather than specific
joint/limb motion.
Addressing these challenges, we develop and validate a new method for fidgety
movement assessment from consumer-grade videos using human poses extracted from
short clips. Human poses capture only relevant motion profiles of joints and
limbs and are thus free from irrelevant appearance artifacts. The dynamics and
coordination between joints are modeled using spatio-temporal graph
convolutional networks. Frames and body parts that contain discriminative
information about fidgety movements are selected through a spatio-temporal
attention mechanism. We validate the proposed model on the cerebral palsy
screening task using a real-life consumer-grade video dataset collected at an
Australian hospital through the Cerebral Palsy Alliance, Australia. Our
experiments show that the proposed method achieves the ROC-AUC score of 81.87%,
significantly outperforming existing competing methods with better
interpretability.
Related papers
- Advanced Gesture Recognition in Autism: Integrating YOLOv7, Video Augmentation and VideoMAE for Video Analysis [9.162792034193373]
This research work aims to identify repetitive behaviors indicative of autism by analyzing videos captured in natural settings as children engage in daily activities.
The focus is on accurately categorizing real-time repetitive gestures such as spinning, head banging, and arm flapping.
A key component of the proposed methodology is the use of textbfVideoMAE, a model designed to improve both spatial and temporal analysis of video data.
arXiv Detail & Related papers (2024-10-12T02:55:37Z) - COIN: Control-Inpainting Diffusion Prior for Human and Camera Motion Estimation [98.05046790227561]
COIN is a control-inpainting motion diffusion prior that enables fine-grained control to disentangle human and camera motions.
COIN outperforms the state-of-the-art methods in terms of global human motion estimation and camera motion estimation.
arXiv Detail & Related papers (2024-08-29T10:36:29Z) - SpaER: Learning Spatio-temporal Equivariant Representations for Fetal Brain Motion Tracking [6.417960463128722]
SpaER is a pioneering method for fetal motion tracking.
We develop an equivariant neural network that efficiently learns rigid motion sequences.
We validate our model using real fetal echo-planar images with simulated and real motions.
arXiv Detail & Related papers (2024-07-29T17:24:52Z) - Dynamic Gaussian Splatting from Markerless Motion Capture can
Reconstruct Infants Movements [2.44755919161855]
This work paves the way for advanced movement analysis tools that can be applied to diverse clinical populations.
We explored the application of dynamic Gaussian splatting to sparse markerless motion capture data.
Our results demonstrate the potential of this method in rendering novel views of scenes and tracking infant movements.
arXiv Detail & Related papers (2023-10-30T11:09:39Z) - Intelligent Knee Sleeves: A Real-time Multimodal Dataset for 3D Lower
Body Motion Estimation Using Smart Textile [2.2008680042670123]
We present a multimodal dataset with benchmarks collected using a novel pair of Intelligent Knee Sleeves for human pose estimation.
Our system utilizes synchronized datasets that comprise time-series data from the Knee Sleeves and the corresponding ground truth labels from the visualized motion capture camera system.
We employ these to generate 3D human models solely based on the wearable data of individuals performing different activities.
arXiv Detail & Related papers (2023-10-02T00:34:21Z) - Vision-Based Activity Recognition in Children with Autism-Related
Behaviors [15.915410623440874]
We demonstrate the effect of a region-based computer vision system to help clinicians and parents analyze a child's behavior.
The data is pre-processed by detecting the target child in the video to reduce the impact of background noise.
Motivated by the effectiveness of temporal convolutional models, we propose both light-weight and conventional models capable of extracting action features from video frames.
arXiv Detail & Related papers (2022-08-08T15:12:27Z) - Investigating Pose Representations and Motion Contexts Modeling for 3D
Motion Prediction [63.62263239934777]
We conduct an indepth study on various pose representations with a focus on their effects on the motion prediction task.
We propose a novel RNN architecture termed AHMR (Attentive Hierarchical Motion Recurrent network) for motion prediction.
Our approach outperforms the state-of-the-art methods in short-term prediction and achieves much enhanced long-term prediction proficiency.
arXiv Detail & Related papers (2021-12-30T10:45:22Z) - Contact-Aware Retargeting of Skinned Motion [49.71236739408685]
This paper introduces a motion estimation method that preserves self-contacts and prevents interpenetration.
The method identifies self-contacts and ground contacts in the input motion, and optimize the motion to apply to the output skeleton.
In experiments, our results quantitatively outperform previous methods and we conduct a user study where our retargeted motions are rated as higher-quality than those produced by recent works.
arXiv Detail & Related papers (2021-09-15T17:05:02Z) - Multi-level Motion Attention for Human Motion Prediction [132.29963836262394]
We study the use of different types of attention, computed at joint, body part, and full pose levels.
Our experiments on Human3.6M, AMASS and 3DPW validate the benefits of our approach for both periodical and non-periodical actions.
arXiv Detail & Related papers (2021-06-17T08:08:11Z) - A Novel Approach for Correcting Multiple Discrete Rigid In-Plane Motions
Artefacts in MRI Scans [63.28835187934139]
We propose a novel method for removing motion artefacts using a deep neural network with two input branches.
The proposed method can be applied to artefacts generated by multiple movements of the patient.
arXiv Detail & Related papers (2020-06-24T15:25:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.