Real-time Pose and Shape Reconstruction of Two Interacting Hands With a
Single Depth Camera
- URL: http://arxiv.org/abs/2106.08059v1
- Date: Tue, 15 Jun 2021 11:39:49 GMT
- Title: Real-time Pose and Shape Reconstruction of Two Interacting Hands With a
Single Depth Camera
- Authors: Franziska Mueller, Micah Davis, Florian Bernard, Oleksandr
Sotnychenko, Mickeal Verschoor, Miguel A. Otaduy, Dan Casas, Christian
Theobalt
- Abstract summary: We present a novel method for real-time pose and shape reconstruction of two strongly interacting hands.
Our approach combines an extensive list of favorable properties, namely it is marker-less.
We show state-of-the-art results in scenes that exceed the complexity level demonstrated by previous work.
- Score: 79.41374930171469
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present a novel method for real-time pose and shape reconstruction of two
strongly interacting hands. Our approach is the first two-hand tracking
solution that combines an extensive list of favorable properties, namely it is
marker-less, uses a single consumer-level depth camera, runs in real time,
handles inter- and intra-hand collisions, and automatically adjusts to the
user's hand shape. In order to achieve this, we embed a recent parametric hand
pose and shape model and a dense correspondence predictor based on a deep
neural network into a suitable energy minimization framework. For training the
correspondence prediction network, we synthesize a two-hand dataset based on
physical simulations that includes both hand pose and shape annotations while
at the same time avoiding inter-hand penetrations. To achieve real-time rates,
we phrase the model fitting in terms of a nonlinear least-squares problem so
that the energy can be optimized based on a highly efficient GPU-based
Gauss-Newton optimizer. We show state-of-the-art results in scenes that exceed
the complexity level demonstrated by previous work, including tight two-hand
grasps, significant inter-hand occlusions, and gesture interaction.
Related papers
- DICE: End-to-end Deformation Capture of Hand-Face Interactions from a Single Image [98.29284902879652]
We present DICE, the first end-to-end method for Deformation-aware hand-face Interaction reCovEry from a single image.
It features disentangling the regression of local deformation fields and global mesh locations into two network branches.
It achieves state-of-the-art performance on a standard benchmark and in-the-wild data in terms of accuracy and physical plausibility.
arXiv Detail & Related papers (2024-06-26T00:08:29Z) - Lightweight Estimation of Hand Mesh and Biomechanically Feasible
Kinematic Parameters [9.477719717840683]
We propose an efficient variation of the previously proposed image-to-lixel approach to efficiently estimate hand meshes from the images.
We introduce an inverted kinematic(IK) network to translate the estimated hand mesh to a biomechanically feasible set of joint rotation parameters.
Our Lite I2L Mesh Net achieves state-of-the-art joint and mesh estimation accuracy with less than $13%$ of the total computational complexity.
arXiv Detail & Related papers (2023-03-26T22:24:12Z) - HandNeRF: Neural Radiance Fields for Animatable Interacting Hands [122.32855646927013]
We propose a novel framework to reconstruct accurate appearance and geometry with neural radiance fields (NeRF) for interacting hands.
We conduct extensive experiments to verify the merits of our proposed HandNeRF and report a series of state-of-the-art results.
arXiv Detail & Related papers (2023-03-24T06:19:19Z) - Im2Hands: Learning Attentive Implicit Representation of Interacting
Two-Hand Shapes [58.551154822792284]
Implicit Two Hands (Im2Hands) is the first neural implicit representation of two interacting hands.
Im2Hands can produce fine-grained geometry of two hands with high hand-to-hand and hand-to-image coherency.
We experimentally demonstrate the effectiveness of Im2Hands on two-hand reconstruction in comparison to related methods.
arXiv Detail & Related papers (2023-02-28T06:38:25Z) - Tracking and Reconstructing Hand Object Interactions from Point Cloud
Sequences in the Wild [35.55753131098285]
We propose a point cloud based hand joint tracking network, HandTrackNet, to estimate the inter-frame hand joint motion.
Our pipeline then reconstructs the full hand via converting the predicted hand joints into a template-based parametric hand model MANO.
For object tracking, we devise a simple yet effective module that estimates the object SDF from the first frame and performs optimization-based tracking.
arXiv Detail & Related papers (2022-09-24T13:40:09Z) - Monocular 3D Reconstruction of Interacting Hands via Collision-Aware
Factorized Refinements [96.40125818594952]
We make the first attempt to reconstruct 3D interacting hands from monocular single RGB images.
Our method can generate 3D hand meshes with both precise 3D poses and minimal collisions.
arXiv Detail & Related papers (2021-11-01T08:24:10Z) - NeuroMorph: Unsupervised Shape Interpolation and Correspondence in One
Go [109.88509362837475]
We present NeuroMorph, a new neural network architecture that takes as input two 3D shapes.
NeuroMorph produces smooth and point-to-point correspondences between them.
It works well for a large variety of input shapes, including non-isometric pairs from different object categories.
arXiv Detail & Related papers (2021-06-17T12:25:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.