Beyond 3DMM: Learning to Capture High-fidelity 3D Face Shape
- URL: http://arxiv.org/abs/2204.04379v1
- Date: Sat, 9 Apr 2022 03:46:18 GMT
- Title: Beyond 3DMM: Learning to Capture High-fidelity 3D Face Shape
- Authors: Xiangyu Zhu, Chang Yu, Di Huang, Zhen Lei, Hao Wang, Stan Z. Li
- Abstract summary: 3D Morphable Model (3DMM) fitting has widely benefited face analysis due to its strong 3D priori.
Previous reconstructed 3D faces suffer from degraded visual verisimilitude due to the loss of fine-grained geometry.
This paper proposes a complete solution to capture the personalized shape so that the reconstructed shape looks identical to the corresponding person.
- Score: 77.95154911528365
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: 3D Morphable Model (3DMM) fitting has widely benefited face analysis due to
its strong 3D priori. However, previous reconstructed 3D faces suffer from
degraded visual verisimilitude due to the loss of fine-grained geometry, which
is attributed to insufficient ground-truth 3D shapes, unreliable training
strategies and limited representation power of 3DMM. To alleviate this issue,
this paper proposes a complete solution to capture the personalized shape so
that the reconstructed shape looks identical to the corresponding person.
Specifically, given a 2D image as the input, we virtually render the image in
several calibrated views to normalize pose variations while preserving the
original image geometry. A many-to-one hourglass network serves as the
encode-decoder to fuse multiview features and generate vertex displacements as
the fine-grained geometry. Besides, the neural network is trained by directly
optimizing the visual effect, where two 3D shapes are compared by measuring the
similarity between the multiview images rendered from the shapes. Finally, we
propose to generate the ground-truth 3D shapes by registering RGB-D images
followed by pose and shape augmentation, providing sufficient data for network
training. Experiments on several challenging protocols demonstrate the superior
reconstruction accuracy of our proposal on the face shape.
Related papers
- 3Deformer: A Common Framework for Image-Guided Mesh Deformation [27.732389685912214]
Given a source 3D mesh with semantic materials, and a user-specified semantic image, 3Deformer can accurately edit the source mesh.
Our 3Deformer is able to produce impressive results and reaches the state-of-the-art level.
arXiv Detail & Related papers (2023-07-19T10:44:44Z) - RAFaRe: Learning Robust and Accurate Non-parametric 3D Face
Reconstruction from Pseudo 2D&3D Pairs [13.11105614044699]
We propose a robust and accurate non-parametric method for single-view 3D face reconstruction (SVFR)
A large-scale pseudo 2D&3D dataset is created by first rendering the detailed 3D faces, then swapping the face in the wild images with the rendered face.
Our model outperforms previous methods on FaceScape-wild/lab and MICC benchmarks.
arXiv Detail & Related papers (2023-02-10T19:40:26Z) - Self-Supervised Geometry-Aware Encoder for Style-Based 3D GAN Inversion [115.82306502822412]
StyleGAN has achieved great progress in 2D face reconstruction and semantic editing via image inversion and latent editing.
A corresponding generic 3D GAN inversion framework is still missing, limiting the applications of 3D face reconstruction and semantic editing.
We study the challenging problem of 3D GAN inversion where a latent code is predicted given a single face image to faithfully recover its 3D shapes and detailed textures.
arXiv Detail & Related papers (2022-12-14T18:49:50Z) - MvDeCor: Multi-view Dense Correspondence Learning for Fine-grained 3D
Segmentation [91.6658845016214]
We propose to utilize self-supervised techniques in the 2D domain for fine-grained 3D shape segmentation tasks.
We render a 3D shape from multiple views, and set up a dense correspondence learning task within the contrastive learning framework.
As a result, the learned 2D representations are view-invariant and geometrically consistent.
arXiv Detail & Related papers (2022-08-18T00:48:15Z) - Pixel2Mesh++: 3D Mesh Generation and Refinement from Multi-View Images [82.32776379815712]
We study the problem of shape generation in 3D mesh representation from a small number of color images with or without camera poses.
We adopt to further improve the shape quality by leveraging cross-view information with a graph convolution network.
Our model is robust to the quality of the initial mesh and the error of camera pose, and can be combined with a differentiable function for test-time optimization.
arXiv Detail & Related papers (2022-04-21T03:42:31Z) - 3D Shape Reconstruction from 2D Images with Disentangled Attribute Flow [61.62796058294777]
Reconstructing 3D shape from a single 2D image is a challenging task.
Most of the previous methods still struggle to extract semantic attributes for 3D reconstruction task.
We propose 3DAttriFlow to disentangle and extract semantic attributes through different semantic levels in the input images.
arXiv Detail & Related papers (2022-03-29T02:03:31Z) - Hard Example Generation by Texture Synthesis for Cross-domain Shape
Similarity Learning [97.56893524594703]
Image-based 3D shape retrieval (IBSR) aims to find the corresponding 3D shape of a given 2D image from a large 3D shape database.
metric learning with some adaptation techniques seems to be a natural solution to shape similarity learning.
We develop a geometry-focused multi-view metric learning framework empowered by texture synthesis.
arXiv Detail & Related papers (2020-10-23T08:52:00Z) - Adaptive 3D Face Reconstruction from a Single Image [45.736818498242016]
We propose a novel joint 2D and 3D optimization method to adaptively reconstruct 3D face shapes from a single image.
Experimental results on multiple datasets demonstrate that our method can generate high-quality reconstruction from a single color image.
arXiv Detail & Related papers (2020-07-08T09:35:26Z) - Single-View 3D Object Reconstruction from Shape Priors in Memory [15.641803721287628]
Existing methods for single-view 3D object reconstruction do not contain enough information to reconstruct high-quality 3D shapes.
We propose a novel method, named Mem3D, that explicitly constructs shape priors to supplement the missing information in the image.
We also propose a voxel triplet loss function that helps to retrieve the precise 3D shapes that are highly related to the input image from shape priors.
arXiv Detail & Related papers (2020-03-08T03:51:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.