Trajectory-guided Motion Perception for Facial Expression Quality Assessment in Neurological Disorders
- URL: http://arxiv.org/abs/2504.09530v2
- Date: Wed, 16 Apr 2025 13:59:54 GMT
- Title: Trajectory-guided Motion Perception for Facial Expression Quality Assessment in Neurological Disorders
- Authors: Shuchao Duan, Amirhossein Dadashzadeh, Alan Whone, Majid Mirmehdi,
- Abstract summary: We introduce Trajectory-guided Motion Perception Transformer (TraMP-Former)<n>TraMP-Former fuses landmark trajectory features for fine-grained motion capture with visual semantic cues from RGB frames, ultimately regressing the combined features into a quality score.<n>Experiments demonstrate that TraMP-Former achieves new state-of-the-art performance on benchmark datasets with neurological disorders.
- Score: 5.169094293336516
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Automated facial expression quality assessment (FEQA) in neurological disorders is critical for enhancing diagnostic accuracy and improving patient care, yet effectively capturing the subtle motions and nuances of facial muscle movements remains a challenge. We propose to analyse facial landmark trajectories, a compact yet informative representation, that encodes these subtle motions from a high-level structural perspective. Hence, we introduce Trajectory-guided Motion Perception Transformer (TraMP-Former), a novel FEQA framework that fuses landmark trajectory features for fine-grained motion capture with visual semantic cues from RGB frames, ultimately regressing the combined features into a quality score. Extensive experiments demonstrate that TraMP-Former achieves new state-of-the-art performance on benchmark datasets with neurological disorders, including PFED5 (up by 6.51%) and an augmented Toronto NeuroFace (up by 7.62%). Our ablation studies further validate the efficiency and effectiveness of landmark trajectories in FEQA. Our code is available at https://github.com/shuchaoduan/TraMP-Former.
Related papers
- Validation of Human Pose Estimation and Human Mesh Recovery for Extracting Clinically Relevant Motion Data from Videos [79.62407455005561]
Marker-less motion capture using human pose estimation produces results in-line with the results of both the IMU and MoCap kinematics.<n>While there is still room for improvement when it comes to the quality of the data produced, we believe that this compromise is within the room of error.
arXiv Detail & Related papers (2025-03-18T22:18:33Z) - Markerless Tracking-Based Registration for Medical Image Motion Correction [0.4288177321445912]
This study focuses on isolating swallowing dynamics from interfering patient motion in videofluoroscopy.<n> Optical flow methods fail due to artifacts like flickering and instability, making them unreliable for distinguishing different motion groups.<n>We introduce a novel motion correction pipeline that effectively removes disruptive motion while preserving swallowing dynamics.
arXiv Detail & Related papers (2025-03-13T11:18:50Z) - FacialPulse: An Efficient RNN-based Depression Detection via Temporal Facial Landmarks [21.076600109388394]
Depression is a prevalent mental health disorder that significantly impacts individuals' lives and well-being.
Recently, there are many end-to-end deep learning methods leveraging the facial expression features for automatic depression detection.
We propose a novel framework called FacialPulse, which recognizes depression with high accuracy and speed.
arXiv Detail & Related papers (2024-08-07T01:50:34Z) - REST: Efficient and Accelerated EEG Seizure Analysis through Residual State Updates [54.96885726053036]
This paper introduces a novel graph-based residual state update mechanism (REST) for real-time EEG signal analysis.
By leveraging a combination of graph neural networks and recurrent structures, REST efficiently captures both non-Euclidean geometry and temporal dependencies within EEG data.
Our model demonstrates high accuracy in both seizure detection and classification tasks.
arXiv Detail & Related papers (2024-06-03T16:30:19Z) - Unsupervised Skin Feature Tracking with Deep Neural Networks [0.0]
Deep convolutional neural networks have shown remarkable accuracy in tracking tasks.
Our pipeline employs a convolutional stacked autoencoder to match image crops with a reference crop containing the target feature.
Our unsupervised learning approach excels in tracking various skin features under significant motion conditions.
arXiv Detail & Related papers (2024-05-08T10:27:05Z) - Joint fMRI Decoding and Encoding with Latent Embedding Alignment [77.66508125297754]
We introduce a unified framework that addresses both fMRI decoding and encoding.
Our model concurrently recovers visual stimuli from fMRI signals and predicts brain activity from images within a unified framework.
arXiv Detail & Related papers (2023-03-26T14:14:58Z) - Attentive Symmetric Autoencoder for Brain MRI Segmentation [56.02577247523737]
We propose a novel Attentive Symmetric Auto-encoder based on Vision Transformer (ViT) for 3D brain MRI segmentation tasks.
In the pre-training stage, the proposed auto-encoder pays more attention to reconstruct the informative patches according to the gradient metrics.
Experimental results show that our proposed attentive symmetric auto-encoder outperforms the state-of-the-art self-supervised learning methods and medical image segmentation models.
arXiv Detail & Related papers (2022-09-19T09:43:19Z) - Neural Motion Fields: Encoding Grasp Trajectories as Implicit Value
Functions [65.84090965167535]
We present Neural Motion Fields, a novel object representation which encodes both object point clouds and the relative task trajectories as an implicit value function parameterized by a neural network.
This object-centric representation models a continuous distribution over the SE(3) space and allows us to perform grasping reactively by leveraging sampling-based MPC to optimize this value function.
arXiv Detail & Related papers (2022-06-29T18:47:05Z) - Neural Moving Horizon Estimation for Robust Flight Control [6.023276947115864]
Estimating and reacting to external disturbances is crucial for robust flight control of quadrotors.
We propose a neural moving horizon estimator (NeuroMHE) that can automatically tune the MHE parameters modeled by a neural network.
NeuroMHE outperforms the state-of-the-art estimator with force estimation error reductions of up to 49.4%.
arXiv Detail & Related papers (2022-06-21T13:43:24Z) - Unsupervised inter-frame motion correction for whole-body dynamic PET
using convolutional long short-term memory in a convolutional neural network [9.349668170221975]
We develop an unsupervised deep learning-based framework to correct inter-frame body motion.
The motion estimation network is a convolutional neural network with a combined convolutional long short-term memory layer.
Once trained, the motion estimation inference time of our proposed network was around 460 times faster than the conventional registration baseline.
arXiv Detail & Related papers (2022-06-13T17:38:16Z) - A Robust Backpropagation-Free Framework for Images [47.97322346441165]
We present an error kernel driven activation alignment algorithm for image data.
EKDAA accomplishes through the introduction of locally derived error transmission kernels and error maps.
Results are presented for an EKDAA trained CNN that employs a non-differentiable activation function.
arXiv Detail & Related papers (2022-06-03T21:14:10Z) - Unsupervised Motion Representation Learning with Capsule Autoencoders [54.81628825371412]
Motion Capsule Autoencoder (MCAE) models motion in a two-level hierarchy.
MCAE is evaluated on a novel Trajectory20 motion dataset and various real-world skeleton-based human action datasets.
arXiv Detail & Related papers (2021-10-01T16:52:03Z) - The FaceChannel: A Fast & Furious Deep Neural Network for Facial
Expression Recognition [71.24825724518847]
Current state-of-the-art models for automatic Facial Expression Recognition (FER) are based on very deep neural networks that are effective but rather expensive to train.
We formalize the FaceChannel, a light-weight neural network that has much fewer parameters than common deep neural networks.
We demonstrate how our model achieves a comparable, if not better, performance to the current state-of-the-art in FER.
arXiv Detail & Related papers (2020-09-15T09:25:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.