CrossHuman: Learning Cross-Guidance from Multi-Frame Images for Human
Reconstruction
- URL: http://arxiv.org/abs/2207.09735v1
- Date: Wed, 20 Jul 2022 08:25:20 GMT
- Title: CrossHuman: Learning Cross-Guidance from Multi-Frame Images for Human
Reconstruction
- Authors: Liliang Chen, Jiaqi Li, Han Huang, Yandong Guo
- Abstract summary: CrossHuman is a novel method that learns cross-guidance from parametric human model and multi-frame RGB images.
We design a reconstruction pipeline combined with tracking-based methods and tracking-free methods.
Compared with previous works, our CrossHuman enables high-fidelity geometry details and texture in both visible and invisible regions.
- Score: 6.450579406495884
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose CrossHuman, a novel method that learns cross-guidance from
parametric human model and multi-frame RGB images to achieve high-quality 3D
human reconstruction. To recover geometry details and texture even in invisible
regions, we design a reconstruction pipeline combined with tracking-based
methods and tracking-free methods. Given a monocular RGB sequence, we track the
parametric human model in the whole sequence, the points (voxels) corresponding
to the target frame are warped to reference frames by the parametric body
motion. Guided by the geometry priors of the parametric body and spatially
aligned features from RGB sequence, the robust implicit surface is fused.
Moreover, a multi-frame transformer (MFT) and a self-supervised warp refinement
module are integrated to the framework to relax the requirements of parametric
body and help to deal with very loose cloth. Compared with previous works, our
CrossHuman enables high-fidelity geometry details and texture in both visible
and invisible regions and improves the accuracy of the human reconstruction
even under estimated inaccurate parametric human models. The experiments
demonstrate that our method achieves state-of-the-art (SOTA) performance.
Related papers
- ANIM: Accurate Neural Implicit Model for Human Reconstruction from a single RGB-D image [40.03212588672639]
ANIM is a novel method that reconstructs arbitrary 3D human shapes from single-view RGB-D images with an unprecedented level of accuracy.
Our model learns geometric details from both pixel-aligned and voxel-aligned features to leverage depth information.
Experiments demonstrate that ANIM outperforms state-of-the-art works that use RGB, surface normals, point cloud or RGB-D data as input.
arXiv Detail & Related papers (2024-03-15T14:45:38Z) - FrozenRecon: Pose-free 3D Scene Reconstruction with Frozen Depth Models [67.96827539201071]
We propose a novel test-time optimization approach for 3D scene reconstruction.
Our method achieves state-of-the-art cross-dataset reconstruction on five zero-shot testing datasets.
arXiv Detail & Related papers (2023-08-10T17:55:02Z) - Towards Scalable Multi-View Reconstruction of Geometry and Materials [27.660389147094715]
We propose a novel method for joint recovery of camera pose, object geometry and spatially-varying Bidirectional Reflectance Distribution Function (svBRDF) of 3D scenes.
The input are high-resolution RGBD images captured by a mobile, hand-held capture system with point lights for active illumination.
arXiv Detail & Related papers (2023-06-06T15:07:39Z) - THUNDR: Transformer-based 3D HUmaN Reconstruction with Markers [67.8628917474705]
THUNDR is a transformer-based deep neural network methodology to reconstruct the 3d pose and shape of people.
We show state-of-the-art results on Human3.6M and 3DPW, for both the fully-supervised and the self-supervised models.
We observe very solid 3d reconstruction performance for difficult human poses collected in the wild.
arXiv Detail & Related papers (2021-06-17T09:09:24Z) - Nothing But Geometric Constraints: A Model-Free Method for Articulated
Object Pose Estimation [89.82169646672872]
We propose an unsupervised vision-based system to estimate the joint configurations of the robot arm from a sequence of RGB or RGB-D images without knowing the model a priori.
We combine a classical geometric formulation with deep learning and extend the use of epipolar multi-rigid-body constraints to solve this task.
arXiv Detail & Related papers (2020-11-30T20:46:48Z) - Neural Descent for Visual 3D Human Pose and Shape [67.01050349629053]
We present deep neural network methodology to reconstruct the 3d pose and shape of people, given an input RGB image.
We rely on a recently introduced, expressivefull body statistical 3d human model, GHUM, trained end-to-end.
Central to our methodology, is a learning to learn and optimize approach, referred to as HUmanNeural Descent (HUND), which avoids both second-order differentiation.
arXiv Detail & Related papers (2020-08-16T13:38:41Z) - TexMesh: Reconstructing Detailed Human Texture and Geometry from RGB-D
Video [37.33902000401107]
TexMesh is a novel approach to reconstruct detailed human meshes with high-resolution full-body texture from RGB-D video.
In practice, we train our models on a short example sequence for self-adaptation and the model runs at interactive framerate afterwards.
arXiv Detail & Related papers (2020-08-01T03:17:23Z) - PaMIR: Parametric Model-Conditioned Implicit Representation for
Image-based Human Reconstruction [67.08350202974434]
We propose Parametric Model-Conditioned Implicit Representation (PaMIR), which combines the parametric body model with the free-form deep implicit function.
We show that our method achieves state-of-the-art performance for image-based 3D human reconstruction in the cases of challenging poses and clothing types.
arXiv Detail & Related papers (2020-07-08T02:26:19Z) - SparseFusion: Dynamic Human Avatar Modeling from Sparse RGBD Images [49.52782544649703]
We propose a novel approach to reconstruct 3D human body shapes based on a sparse set of RGBD frames.
The main challenge is how to robustly fuse these sparse frames into a canonical 3D model.
Our framework is flexible, with potential applications going beyond shape reconstruction.
arXiv Detail & Related papers (2020-06-05T18:53:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.