Decoupled Iterative Refinement Framework for Interacting Hands
Reconstruction from a Single RGB Image
- URL: http://arxiv.org/abs/2302.02410v2
- Date: Mon, 21 Aug 2023 03:46:50 GMT
- Title: Decoupled Iterative Refinement Framework for Interacting Hands
Reconstruction from a Single RGB Image
- Authors: Pengfei Ren, Chao Wen, Xiaozheng Zheng, Zhou Xue, Haifeng Sun, Qi Qi,
Jingyu Wang, Jianxin Liao
- Abstract summary: We propose a decoupled iterative refinement framework to achieve pixel-alignment hand reconstruction.
Our method outperforms all existing two-hand reconstruction methods by a large margin on the InterHand2.6M dataset.
- Score: 30.24438569170251
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Reconstructing interacting hands from a single RGB image is a very
challenging task. On the one hand, severe mutual occlusion and similar local
appearance between two hands confuse the extraction of visual features,
resulting in the misalignment of estimated hand meshes and the image. On the
other hand, there are complex spatial relationship between interacting hands,
which significantly increases the solution space of hand poses and increases
the difficulty of network learning. In this paper, we propose a decoupled
iterative refinement framework to achieve pixel-alignment hand reconstruction
while efficiently modeling the spatial relationship between hands.
Specifically, we define two feature spaces with different characteristics,
namely 2D visual feature space and 3D joint feature space. First, we obtain
joint-wise features from the visual feature map and utilize a graph convolution
network and a transformer to perform intra- and inter-hand information
interaction in the 3D joint feature space, respectively. Then, we project the
joint features with global information back into the 2D visual feature space in
an obfuscation-free manner and utilize the 2D convolution for pixel-wise
enhancement. By performing multiple alternate enhancements in the two feature
spaces, our method can achieve an accurate and robust reconstruction of
interacting hands. Our method outperforms all existing two-hand reconstruction
methods by a large margin on the InterHand2.6M dataset.
Related papers
- BIFRÖST: 3D-Aware Image compositing with Language Instructions [27.484947109237964]
Bifr"ost is a novel 3D-aware framework that is built upon diffusion models to perform instruction-based image composition.
Bifr"ost addresses issues by training MLLM as a 2.5D location predictor and integrating depth maps as an extra condition during the generation process.
arXiv Detail & Related papers (2024-10-24T18:35:12Z) - Learning Interaction-aware 3D Gaussian Splatting for One-shot Hand Avatars [47.61442517627826]
We propose to create animatable avatars for interacting hands with 3D Gaussian Splatting (GS) and single-image inputs.
Our proposed method is validated via extensive experiments on the large-scale InterHand2.6M dataset.
arXiv Detail & Related papers (2024-10-11T14:14:51Z) - DICE: End-to-end Deformation Capture of Hand-Face Interactions from a Single Image [98.29284902879652]
We present DICE, the first end-to-end method for Deformation-aware hand-face Interaction reCovEry from a single image.
It features disentangling the regression of local deformation fields and global mesh locations into two network branches.
It achieves state-of-the-art performance on a standard benchmark and in-the-wild data in terms of accuracy and physical plausibility.
arXiv Detail & Related papers (2024-06-26T00:08:29Z) - HandNeRF: Neural Radiance Fields for Animatable Interacting Hands [122.32855646927013]
We propose a novel framework to reconstruct accurate appearance and geometry with neural radiance fields (NeRF) for interacting hands.
We conduct extensive experiments to verify the merits of our proposed HandNeRF and report a series of state-of-the-art results.
arXiv Detail & Related papers (2023-03-24T06:19:19Z) - Im2Hands: Learning Attentive Implicit Representation of Interacting
Two-Hand Shapes [58.551154822792284]
Implicit Two Hands (Im2Hands) is the first neural implicit representation of two interacting hands.
Im2Hands can produce fine-grained geometry of two hands with high hand-to-hand and hand-to-image coherency.
We experimentally demonstrate the effectiveness of Im2Hands on two-hand reconstruction in comparison to related methods.
arXiv Detail & Related papers (2023-02-28T06:38:25Z) - Monocular 3D Reconstruction of Interacting Hands via Collision-Aware
Factorized Refinements [96.40125818594952]
We make the first attempt to reconstruct 3D interacting hands from monocular single RGB images.
Our method can generate 3D hand meshes with both precise 3D poses and minimal collisions.
arXiv Detail & Related papers (2021-11-01T08:24:10Z) - RGB2Hands: Real-Time Tracking of 3D Hand Interactions from Monocular RGB
Video [76.86512780916827]
We present the first real-time method for motion capture of skeletal pose and 3D surface geometry of hands from a single RGB camera.
In order to address the inherent depth ambiguities in RGB data, we propose a novel multi-task CNN.
We experimentally verify the individual components of our RGB two-hand tracking and 3D reconstruction pipeline.
arXiv Detail & Related papers (2021-06-22T12:53:56Z) - A hybrid classification-regression approach for 3D hand pose estimation
using graph convolutional networks [1.0152838128195467]
We propose a two-stage GCN-based framework that learns per-pose relationship constraints.
The first phase quantizes the 2D/3D space to classify the joints into 2D/3D blocks based on their locality.
The second stage uses a GCN-based module that uses an adaptative nearest neighbor algorithm to determine joint relationships.
arXiv Detail & Related papers (2021-05-23T10:09:10Z) - Joint Hand-object 3D Reconstruction from a Single Image with
Cross-branch Feature Fusion [78.98074380040838]
We propose to consider hand and object jointly in feature space and explore the reciprocity of the two branches.
We employ an auxiliary depth estimation module to augment the input RGB image with the estimated depth map.
Our approach significantly outperforms existing approaches in terms of the reconstruction accuracy of objects.
arXiv Detail & Related papers (2020-06-28T09:50:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.