Robustness Evaluation in Hand Pose Estimation Models using Metamorphic
Testing
- URL: http://arxiv.org/abs/2303.04566v1
- Date: Wed, 8 Mar 2023 13:23:53 GMT
- Title: Robustness Evaluation in Hand Pose Estimation Models using Metamorphic
Testing
- Authors: Muxin Pu, Chun Yong Chong, Mei Kuan Lim
- Abstract summary: Hand pose estimation (HPE) is a task that predicts and describes the hand poses from images or video frames.
In this work, we adopt metamorphic testing to evaluate the robustness of HPE models.
- Score: 2.535271349350579
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Hand pose estimation (HPE) is a task that predicts and describes the hand
poses from images or video frames. When HPE models estimate hand poses captured
in a laboratory or under controlled environments, they normally deliver good
performance. However, the real-world environment is complex, and various
uncertainties may happen, which could degrade the performance of HPE models.
For example, the hands could be occluded, the visibility of hands could be
reduced by imperfect exposure rate, and the contour of hands prone to be
blurred during fast hand movements. In this work, we adopt metamorphic testing
to evaluate the robustness of HPE models and provide suggestions on the choice
of HPE models for different applications. The robustness evaluation was
conducted on four state-of-the-art models, namely MediaPipe hands, OpenPose,
BodyHands, and NSRM hand. We found that on average more than 80\% of the hands
could not be identified by BodyHands, and at least 50\% of hands could not be
identified by MediaPipe hands when diagonal motion blur is introduced, while an
average of more than 50\% of strongly underexposed hands could not be correctly
estimated by NSRM hand. Similarly, applying occlusions on only four hand joints
will also largely degrade the performance of these models. The experimental
results show that occlusions, illumination variations, and motion blur are the
main obstacles to the performance of existing HPE models. These findings may
pave the way for researchers to improve the performance and robustness of hand
pose estimation models and their applications.
Related papers
- HandDiff: 3D Hand Pose Estimation with Diffusion on Image-Point Cloud [60.47544798202017]
Hand pose estimation is a critical task in various human-computer interaction applications.
This paper proposes HandDiff, a diffusion-based hand pose estimation model that iteratively denoises accurate hand pose conditioned on hand-shaped image-point clouds.
Experimental results demonstrate that the proposed HandDiff significantly outperforms the existing approaches on four challenging hand pose benchmark datasets.
arXiv Detail & Related papers (2024-04-04T02:15:16Z) - On the Utility of 3D Hand Poses for Action Recognition [36.64538554919222]
HandFormer is a novel multimodal transformer to efficiently model hand-object interactions.
We factorize hand modeling and represent each joint by its short-term trajectories.
We achieve new state-of-the-art performance on Assembly101 and H2O with significant improvements in egocentric action recognition.
arXiv Detail & Related papers (2024-03-14T18:52:34Z) - Deformer: Dynamic Fusion Transformer for Robust Hand Pose Estimation [59.3035531612715]
Existing methods often struggle to generate plausible hand poses when the hand is heavily occluded or blurred.
In videos, the movements of the hand allow us to observe various parts of the hand that may be occluded or blurred in a single frame.
We propose the Deformer: a framework that implicitly reasons about the relationship between hand parts within the same image.
arXiv Detail & Related papers (2023-03-09T02:24:30Z) - EvHandPose: Event-based 3D Hand Pose Estimation with Sparse Supervision [50.060055525889915]
Event camera shows great potential in 3D hand pose estimation, especially addressing the challenges of fast motion and high dynamic range in a low-power way.
It is challenging to design event representation to encode hand motion information especially when the hands are not moving.
In this paper, we propose EvHandPose with novel hand flow representations in Event-to-Pose module for accurate hand pose estimation.
arXiv Detail & Related papers (2023-03-06T03:27:17Z) - 3D Interacting Hand Pose Estimation by Hand De-occlusion and Removal [85.30756038989057]
Estimating 3D interacting hand pose from a single RGB image is essential for understanding human actions.
We propose to decompose the challenging interacting hand pose estimation task and estimate the pose of each hand separately.
Experiments show that the proposed method significantly outperforms previous state-of-the-art interacting hand pose estimation approaches.
arXiv Detail & Related papers (2022-07-22T13:04:06Z) - Learning to Disambiguate Strongly Interacting Hands via Probabilistic
Per-pixel Part Segmentation [84.28064034301445]
Self-similarity, and the resulting ambiguities in assigning pixel observations to the respective hands, is a major cause of the final 3D pose error.
We propose DIGIT, a novel method for estimating the 3D poses of two interacting hands from a single monocular image.
We experimentally show that the proposed approach achieves new state-of-the-art performance on the InterHand2.6M dataset.
arXiv Detail & Related papers (2021-07-01T13:28:02Z) - Measuring Generalisation to Unseen Viewpoints, Articulations, Shapes and
Objects for 3D Hand Pose Estimation under Hand-Object Interaction [137.28465645405655]
HANDS'19 is a challenge to evaluate the abilities of current 3D hand pose estimators (HPEs) to interpolate and extrapolate the poses of a training set.
We show that the accuracy of state-of-the-art methods can drop, and that they fail mostly on poses absent from the training set.
arXiv Detail & Related papers (2020-03-30T19:28:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.