EvHandPose: Event-based 3D Hand Pose Estimation with Sparse Supervision
- URL: http://arxiv.org/abs/2303.02862v3
- Date: Thu, 28 Dec 2023 08:48:37 GMT
- Title: EvHandPose: Event-based 3D Hand Pose Estimation with Sparse Supervision
- Authors: Jianping Jiang, Jiahe Li, Baowen Zhang, Xiaoming Deng, Boxin Shi
- Abstract summary: Event camera shows great potential in 3D hand pose estimation, especially addressing the challenges of fast motion and high dynamic range in a low-power way.
It is challenging to design event representation to encode hand motion information especially when the hands are not moving.
In this paper, we propose EvHandPose with novel hand flow representations in Event-to-Pose module for accurate hand pose estimation.
- Score: 50.060055525889915
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Event camera shows great potential in 3D hand pose estimation, especially
addressing the challenges of fast motion and high dynamic range in a low-power
way. However, due to the asynchronous differential imaging mechanism, it is
challenging to design event representation to encode hand motion information
especially when the hands are not moving (causing motion ambiguity), and it is
infeasible to fully annotate the temporally dense event stream. In this paper,
we propose EvHandPose with novel hand flow representations in Event-to-Pose
module for accurate hand pose estimation and alleviating the motion ambiguity
issue. To solve the problem under sparse annotation, we design contrast
maximization and hand-edge constraints in Pose-to-IWE (Image with Warped
Events) module and formulate EvHandPose in a weakly-supervision framework. We
further build EvRealHands, the first large-scale real-world event-based hand
pose dataset on several challenging scenes to bridge the real-synthetic domain
gap. Experiments on EvRealHands demonstrate that EvHandPose outperforms
previous event-based methods under all evaluation scenes, achieves accurate and
stable hand pose estimation with high temporal resolution in fast motion and
strong light scenes compared with RGB-based methods, generalizes well to
outdoor scenes and another type of event camera, and shows the potential for
the hand gesture recognition task.
Related papers
- HandDiff: 3D Hand Pose Estimation with Diffusion on Image-Point Cloud [60.47544798202017]
Hand pose estimation is a critical task in various human-computer interaction applications.
This paper proposes HandDiff, a diffusion-based hand pose estimation model that iteratively denoises accurate hand pose conditioned on hand-shaped image-point clouds.
Experimental results demonstrate that the proposed HandDiff significantly outperforms the existing approaches on four challenging hand pose benchmark datasets.
arXiv Detail & Related papers (2024-04-04T02:15:16Z) - On the Utility of 3D Hand Poses for Action Recognition [36.64538554919222]
HandFormer is a novel multimodal transformer to efficiently model hand-object interactions.
We factorize hand modeling and represent each joint by its short-term trajectories.
We achieve new state-of-the-art performance on Assembly101 and H2O with significant improvements in egocentric action recognition.
arXiv Detail & Related papers (2024-03-14T18:52:34Z) - 3D Pose Estimation of Two Interacting Hands from a Monocular Event
Camera [59.846927201816776]
This paper introduces the first framework for 3D tracking of two fast-moving and interacting hands from a single monocular event camera.
Our approach tackles the left-right hand ambiguity with a novel semi-supervised feature-wise attention mechanism and integrates an intersection loss to fix hand collisions.
arXiv Detail & Related papers (2023-12-21T18:59:57Z) - Denoising Diffusion for 3D Hand Pose Estimation from Images [38.20064386142944]
This paper addresses the problem of 3D hand pose estimation from monocular images or sequences.
We present a novel end-to-end framework for 3D hand regression that employs diffusion models that have shown excellent ability to capture the distribution of data for generative purposes.
The proposed model provides state-of-the-art performance when lifting a 2D single-hand image to 3D.
arXiv Detail & Related papers (2023-08-18T12:57:22Z) - Deformer: Dynamic Fusion Transformer for Robust Hand Pose Estimation [59.3035531612715]
Existing methods often struggle to generate plausible hand poses when the hand is heavily occluded or blurred.
In videos, the movements of the hand allow us to observe various parts of the hand that may be occluded or blurred in a single frame.
We propose the Deformer: a framework that implicitly reasons about the relationship between hand parts within the same image.
arXiv Detail & Related papers (2023-03-09T02:24:30Z) - 3D Interacting Hand Pose Estimation by Hand De-occlusion and Removal [85.30756038989057]
Estimating 3D interacting hand pose from a single RGB image is essential for understanding human actions.
We propose to decompose the challenging interacting hand pose estimation task and estimate the pose of each hand separately.
Experiments show that the proposed method significantly outperforms previous state-of-the-art interacting hand pose estimation approaches.
arXiv Detail & Related papers (2022-07-22T13:04:06Z) - Measuring Generalisation to Unseen Viewpoints, Articulations, Shapes and
Objects for 3D Hand Pose Estimation under Hand-Object Interaction [137.28465645405655]
HANDS'19 is a challenge to evaluate the abilities of current 3D hand pose estimators (HPEs) to interpolate and extrapolate the poses of a training set.
We show that the accuracy of state-of-the-art methods can drop, and that they fail mostly on poses absent from the training set.
arXiv Detail & Related papers (2020-03-30T19:28:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.