SeqHAND:RGB-Sequence-Based 3D Hand Pose and Shape Estimation
- URL: http://arxiv.org/abs/2007.05168v1
- Date: Fri, 10 Jul 2020 05:11:14 GMT
- Title: SeqHAND:RGB-Sequence-Based 3D Hand Pose and Shape Estimation
- Authors: John Yang, Hyung Jin Chang, Seungeui Lee, Nojun Kwak
- Abstract summary: 3D hand pose estimation based on RGB images has been studied for a long time.
We propose a novel method that generates a synthetic dataset that mimics natural human hand movements.
We show that utilizing temporal information for 3D hand pose estimation significantly enhances general pose estimations.
- Score: 48.456638103309544
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: 3D hand pose estimation based on RGB images has been studied for a long time.
Most of the studies, however, have performed frame-by-frame estimation based on
independent static images. In this paper, we attempt to not only consider the
appearance of a hand but incorporate the temporal movement information of a
hand in motion into the learning framework for better 3D hand pose estimation
performance, which leads to the necessity of a large scale dataset with
sequential RGB hand images. We propose a novel method that generates a
synthetic dataset that mimics natural human hand movements by re-engineering
annotations of an extant static hand pose dataset into pose-flows. With the
generated dataset, we train a newly proposed recurrent framework, exploiting
visuo-temporal features from sequential images of synthetic hands in motion and
emphasizing temporal smoothness of estimations with a temporal consistency
constraint. Our novel training strategy of detaching the recurrent layer of the
framework during domain finetuning from synthetic to real allows preservation
of the visuo-temporal features learned from sequential synthetic hand images.
Hand poses that are sequentially estimated consequently produce natural and
smooth hand movements which lead to more robust estimations. We show that
utilizing temporal information for 3D hand pose estimation significantly
enhances general pose estimations by outperforming state-of-the-art methods in
experiments on hand pose estimation benchmarks.
Related papers
- HandDiff: 3D Hand Pose Estimation with Diffusion on Image-Point Cloud [60.47544798202017]
Hand pose estimation is a critical task in various human-computer interaction applications.
This paper proposes HandDiff, a diffusion-based hand pose estimation model that iteratively denoises accurate hand pose conditioned on hand-shaped image-point clouds.
Experimental results demonstrate that the proposed HandDiff significantly outperforms the existing approaches on four challenging hand pose benchmark datasets.
arXiv Detail & Related papers (2024-04-04T02:15:16Z) - Mesh Represented Recycle Learning for 3D Hand Pose and Mesh Estimation [3.126179109712709]
We propose a mesh represented recycle learning strategy for 3D hand pose and mesh estimation.
To be specific, a hand pose and mesh estimation model first predicts parametric 3D hand annotations.
Second, synthetic hand images are generated with self-estimated hand mesh representations.
Third, the synthetic hand images are fed into the same model again.
arXiv Detail & Related papers (2023-10-18T09:50:09Z) - Denoising Diffusion for 3D Hand Pose Estimation from Images [38.20064386142944]
This paper addresses the problem of 3D hand pose estimation from monocular images or sequences.
We present a novel end-to-end framework for 3D hand regression that employs diffusion models that have shown excellent ability to capture the distribution of data for generative purposes.
The proposed model provides state-of-the-art performance when lifting a 2D single-hand image to 3D.
arXiv Detail & Related papers (2023-08-18T12:57:22Z) - Exploiting Spatial-Temporal Context for Interacting Hand Reconstruction
on Monocular RGB Video [104.69686024776396]
Reconstructing interacting hands from monocular RGB data is a challenging task, as it involves many interfering factors.
Previous works only leverage information from a single RGB image without modeling their physically plausible relation.
In this work, we are dedicated to explicitly exploiting spatial-temporal information to achieve better interacting hand reconstruction.
arXiv Detail & Related papers (2023-08-08T06:16:37Z) - HandNeRF: Neural Radiance Fields for Animatable Interacting Hands [122.32855646927013]
We propose a novel framework to reconstruct accurate appearance and geometry with neural radiance fields (NeRF) for interacting hands.
We conduct extensive experiments to verify the merits of our proposed HandNeRF and report a series of state-of-the-art results.
arXiv Detail & Related papers (2023-03-24T06:19:19Z) - 3D Interacting Hand Pose Estimation by Hand De-occlusion and Removal [85.30756038989057]
Estimating 3D interacting hand pose from a single RGB image is essential for understanding human actions.
We propose to decompose the challenging interacting hand pose estimation task and estimate the pose of each hand separately.
Experiments show that the proposed method significantly outperforms previous state-of-the-art interacting hand pose estimation approaches.
arXiv Detail & Related papers (2022-07-22T13:04:06Z) - 3D Hand Pose and Shape Estimation from RGB Images for Improved
Keypoint-Based Hand-Gesture Recognition [25.379923604213626]
This paper presents a keypoint-based end-to-end framework for the 3D hand and pose estimation.
It is successfully applied to the hand-gesture recognition task as a study case.
arXiv Detail & Related papers (2021-09-28T17:07:43Z) - Leveraging Photometric Consistency over Time for Sparsely Supervised
Hand-Object Reconstruction [118.21363599332493]
We present a method to leverage photometric consistency across time when annotations are only available for a sparse subset of frames in a video.
Our model is trained end-to-end on color images to jointly reconstruct hands and objects in 3D by inferring their poses.
We achieve state-of-the-art results on 3D hand-object reconstruction benchmarks and demonstrate that our approach allows us to improve the pose estimation accuracy.
arXiv Detail & Related papers (2020-04-28T12:03:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.