Towards 3D Face Reconstruction in Perspective Projection: Estimating
6DoF Face Pose from Monocular Image
- URL: http://arxiv.org/abs/2205.04126v2
- Date: Wed, 17 May 2023 11:35:41 GMT
- Title: Towards 3D Face Reconstruction in Perspective Projection: Estimating
6DoF Face Pose from Monocular Image
- Authors: Yueying Kao and Bowen Pan and Miao Xu and Jiangjing Lyu and Xiangyu
Zhu and Yuanzhang Chang and Xiaobo Li and Zhen Lei
- Abstract summary: In some scenarios that the face is very close to camera or moving along the camera axis, the methods suffer from the inaccurate reconstruction and unstable temporal fitting.
Deep neural network, Perspective Network (PerspNet), is proposed to simultaneously reconstruct 3D face shape in canonical space.
We contribute a large ARKitFace dataset to enable the training and evaluation of 3D face reconstruction solutions under the scenarios of perspective projection.
- Score: 48.77844225075744
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In 3D face reconstruction, orthogonal projection has been widely employed to
substitute perspective projection to simplify the fitting process. This
approximation performs well when the distance between camera and face is far
enough. However, in some scenarios that the face is very close to camera or
moving along the camera axis, the methods suffer from the inaccurate
reconstruction and unstable temporal fitting due to the distortion under the
perspective projection. In this paper, we aim to address the problem of
single-image 3D face reconstruction under perspective projection. Specifically,
a deep neural network, Perspective Network (PerspNet), is proposed to
simultaneously reconstruct 3D face shape in canonical space and learn the
correspondence between 2D pixels and 3D points, by which the 6DoF (6 Degrees of
Freedom) face pose can be estimated to represent perspective projection.
Besides, we contribute a large ARKitFace dataset to enable the training and
evaluation of 3D face reconstruction solutions under the scenarios of
perspective projection, which has 902,724 2D facial images with ground-truth 3D
face mesh and annotated 6DoF pose parameters. Experimental results show that
our approach outperforms current state-of-the-art methods by a significant
margin. The code and data are available at
https://github.com/cbsropenproject/6dof_face.
Related papers
- FrozenRecon: Pose-free 3D Scene Reconstruction with Frozen Depth Models [67.96827539201071]
We propose a novel test-time optimization approach for 3D scene reconstruction.
Our method achieves state-of-the-art cross-dataset reconstruction on five zero-shot testing datasets.
arXiv Detail & Related papers (2023-08-10T17:55:02Z) - RAFaRe: Learning Robust and Accurate Non-parametric 3D Face
Reconstruction from Pseudo 2D&3D Pairs [13.11105614044699]
We propose a robust and accurate non-parametric method for single-view 3D face reconstruction (SVFR)
A large-scale pseudo 2D&3D dataset is created by first rendering the detailed 3D faces, then swapping the face in the wild images with the rendered face.
Our model outperforms previous methods on FaceScape-wild/lab and MICC benchmarks.
arXiv Detail & Related papers (2023-02-10T19:40:26Z) - Perspective Reconstruction of Human Faces by Joint Mesh and Landmark
Regression [89.8129467907451]
We propose to simultaneously reconstruct 3D face mesh in the world space and predict 2D face landmarks on the image plane.
Based on the predicted 3D and 2D landmarks, the 6DoF (6 Degrees Freedom) face pose can be easily estimated by the solver.
arXiv Detail & Related papers (2022-08-15T12:32:20Z) - Neural 3D Scene Reconstruction with the Manhattan-world Assumption [58.90559966227361]
This paper addresses the challenge of reconstructing 3D indoor scenes from multi-view images.
Planar constraints can be conveniently integrated into the recent implicit neural representation-based reconstruction methods.
The proposed method outperforms previous methods by a large margin on 3D reconstruction quality.
arXiv Detail & Related papers (2022-05-05T17:59:55Z) - Beyond 3DMM: Learning to Capture High-fidelity 3D Face Shape [77.95154911528365]
3D Morphable Model (3DMM) fitting has widely benefited face analysis due to its strong 3D priori.
Previous reconstructed 3D faces suffer from degraded visual verisimilitude due to the loss of fine-grained geometry.
This paper proposes a complete solution to capture the personalized shape so that the reconstructed shape looks identical to the corresponding person.
arXiv Detail & Related papers (2022-04-09T03:46:18Z) - FaceScape: 3D Facial Dataset and Benchmark for Single-View 3D Face
Reconstruction [29.920622006999732]
We present a large-scale detailed 3D face dataset, FaceScape, and the corresponding benchmark to evaluate single-view facial 3D reconstruction.
By training on FaceScape data, a novel algorithm is proposed to predict elaborate riggable 3D face models from a single image input.
We also use FaceScape data to generate the in-the-wild and in-the-lab benchmark to evaluate recent methods of single-view face reconstruction.
arXiv Detail & Related papers (2021-11-01T16:48:34Z) - Adaptive 3D Face Reconstruction from a Single Image [45.736818498242016]
We propose a novel joint 2D and 3D optimization method to adaptively reconstruct 3D face shapes from a single image.
Experimental results on multiple datasets demonstrate that our method can generate high-quality reconstruction from a single color image.
arXiv Detail & Related papers (2020-07-08T09:35:26Z) - DeepFaceFlow: In-the-wild Dense 3D Facial Motion Estimation [56.56575063461169]
DeepFaceFlow is a robust, fast, and highly-accurate framework for the estimation of 3D non-rigid facial flow.
Our framework was trained and tested on two very large-scale facial video datasets.
Given registered pairs of images, our framework generates 3D flow maps at 60 fps.
arXiv Detail & Related papers (2020-05-14T23:56:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.