Challenges in Video-Based Infant Action Recognition: A Critical
Examination of the State of the Art
- URL: http://arxiv.org/abs/2311.12300v1
- Date: Tue, 21 Nov 2023 02:36:47 GMT
- Title: Challenges in Video-Based Infant Action Recognition: A Critical
Examination of the State of the Art
- Authors: Elaheh Hatamimajoumerd, Pooria Daneshvar Kakhaki, Xiaofei Huang,
Lingfei Luan, Somaieh Amraee, Sarah Ostadabbas
- Abstract summary: We introduce a groundbreaking dataset called InfActPrimitive'', encompassing five significant infant milestone action categories.
We conduct an extensive comparative analysis employing cutting-edge skeleton-based action recognition models.
Our findings reveal that, although the PoseC3D model achieves the highest accuracy at approximately 71%, the remaining models struggle to accurately capture the dynamics of infant actions.
- Score: 9.327466428403916
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Automated human action recognition, a burgeoning field within computer
vision, boasts diverse applications spanning surveillance, security,
human-computer interaction, tele-health, and sports analysis. Precise action
recognition in infants serves a multitude of pivotal purposes, encompassing
safety monitoring, developmental milestone tracking, early intervention for
developmental delays, fostering parent-infant bonds, advancing computer-aided
diagnostics, and contributing to the scientific comprehension of child
development. This paper delves into the intricacies of infant action
recognition, a domain that has remained relatively uncharted despite the
accomplishments in adult action recognition. In this study, we introduce a
groundbreaking dataset called ``InfActPrimitive'', encompassing five
significant infant milestone action categories, and we incorporate specialized
preprocessing for infant data. We conducted an extensive comparative analysis
employing cutting-edge skeleton-based action recognition models using this
dataset. Our findings reveal that, although the PoseC3D model achieves the
highest accuracy at approximately 71%, the remaining models struggle to
accurately capture the dynamics of infant actions. This highlights a
substantial knowledge gap between infant and adult action recognition domains
and the urgent need for data-efficient pipeline models.
Related papers
- ToddlerAct: A Toddler Action Recognition Dataset for Gross Motor Development Assessment [26.16139407666899]
ToddlerAct is a toddler gross motor action recognition dataset.
We describe the data collection process, annotation methodology, and dataset characteristics.
Our findings highlight the importance of domain-specific datasets for accurate assessment of gross motor development in toddlers.
arXiv Detail & Related papers (2024-08-31T04:31:46Z) - Data Augmentation in Human-Centric Vision [54.97327269866757]
This survey presents a comprehensive analysis of data augmentation techniques in human-centric vision tasks.
It delves into a wide range of research areas including person ReID, human parsing, human pose estimation, and pedestrian detection.
Our work categorizes data augmentation methods into two main types: data generation and data perturbation.
arXiv Detail & Related papers (2024-03-13T16:05:18Z) - Modeling 3D Infant Kinetics Using Adaptive Graph Convolutional Networks [2.2279946664123664]
Spontaneous motor activity, orkinetics', is shown to provide a powerful surrogate measure of upcoming neurodevelopment.
Here, we follow an alternative approach, predicting infants' maturation based on data-driven evaluation of individual motor patterns.
arXiv Detail & Related papers (2024-02-22T09:34:48Z) - Dynamic Gaussian Splatting from Markerless Motion Capture can
Reconstruct Infants Movements [2.44755919161855]
This work paves the way for advanced movement analysis tools that can be applied to diverse clinical populations.
We explored the application of dynamic Gaussian splatting to sparse markerless motion capture data.
Our results demonstrate the potential of this method in rendering novel views of scenes and tracking infant movements.
arXiv Detail & Related papers (2023-10-30T11:09:39Z) - NeuroExplainer: Fine-Grained Attention Decoding to Uncover Cortical
Development Patterns of Preterm Infants [73.85768093666582]
We propose an explainable geometric deep network dubbed NeuroExplainer.
NeuroExplainer is used to uncover altered infant cortical development patterns associated with preterm birth.
arXiv Detail & Related papers (2023-01-01T12:48:12Z) - Language-Assisted Deep Learning for Autistic Behaviors Recognition [13.200025637384897]
We show that a vision-based problem behaviors recognition system can achieve high accuracy and outperform the previous methods by a large margin.
We propose a two-branch multimodal deep learning framework by incorporating the "freely available" language description for each type of problem behavior.
Experimental results demonstrate that incorporating additional language supervision can bring an obvious performance boost for the autism problem behaviors recognition task.
arXiv Detail & Related papers (2022-11-17T02:58:55Z) - Vision-Based Activity Recognition in Children with Autism-Related
Behaviors [15.915410623440874]
We demonstrate the effect of a region-based computer vision system to help clinicians and parents analyze a child's behavior.
The data is pre-processed by detecting the target child in the video to reduce the impact of background noise.
Motivated by the effectiveness of temporal convolutional models, we propose both light-weight and conventional models capable of extracting action features from video frames.
arXiv Detail & Related papers (2022-08-08T15:12:27Z) - Overcoming the Domain Gap in Neural Action Representations [60.47807856873544]
3D pose data can now be reliably extracted from multi-view video sequences without manual intervention.
We propose to use it to guide the encoding of neural action representations together with a set of neural and behavioral augmentations.
To reduce the domain gap, during training, we swap neural and behavioral data across animals that seem to be performing similar actions.
arXiv Detail & Related papers (2021-12-02T12:45:46Z) - Overcoming the Domain Gap in Contrastive Learning of Neural Action
Representations [60.47807856873544]
A fundamental goal in neuroscience is to understand the relationship between neural activity and behavior.
We generated a new multimodal dataset consisting of the spontaneous behaviors generated by fruit flies.
This dataset and our new set of augmentations promise to accelerate the application of self-supervised learning methods in neuroscience.
arXiv Detail & Related papers (2021-11-29T15:27:51Z) - Affect Analysis in-the-wild: Valence-Arousal, Expressions, Action Units
and a Unified Framework [83.21732533130846]
The paper focuses on large in-the-wild databases, i.e., Aff-Wild and Aff-Wild2.
It presents the design of two classes of deep neural networks trained with these databases.
A novel multi-task and holistic framework is presented which is able to jointly learn and effectively generalize and perform affect recognition.
arXiv Detail & Related papers (2021-03-29T17:36:20Z) - A Developmental Neuro-Robotics Approach for Boosting the Recognition of
Handwritten Digits [91.3755431537592]
Recent evidence shows that a simulation of the children's embodied strategies can improve the machine intelligence too.
This article explores the application of embodied strategies to convolutional neural network models in the context of developmental neuro-robotics.
arXiv Detail & Related papers (2020-03-23T14:55:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.