Pixel-Face: A Large-Scale, High-Resolution Benchmark for 3D Face
Reconstruction
- URL: http://arxiv.org/abs/2008.12444v3
- Date: Thu, 3 Sep 2020 02:33:25 GMT
- Title: Pixel-Face: A Large-Scale, High-Resolution Benchmark for 3D Face
Reconstruction
- Authors: Jiangjing Lyu, Xiaobo Li, Xiangyu Zhu, Cheng Cheng
- Abstract summary: We introduce Pixel-Face, a large-scale, high-resolution and diverse 3D face dataset with massive annotations.
Specifically, Pixel-Face contains 855 subjects aging from 18 to 80. Each subject has more than 20 samples with various expressions.
We show that the obtained Pixel-3DM is better in modeling a wide range of face shapes and expressions.
- Score: 15.51331644571456
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: 3D face reconstruction is a fundamental task that can facilitate numerous
applications such as robust facial analysis and augmented reality. It is also a
challenging task due to the lack of high-quality datasets that can fuel current
deep learning-based methods. However, existing datasets are limited in
quantity, realisticity and diversity. To circumvent these hurdles, we introduce
Pixel-Face, a large-scale, high-resolution and diverse 3D face dataset with
massive annotations. Specifically, Pixel-Face contains 855 subjects aging from
18 to 80. Each subject has more than 20 samples with various expressions. Each
sample is composed of high-resolution multi-view RGB images and 3D meshes with
various expressions. Moreover, we collect precise landmarks annotation and 3D
registration result for each data. To demonstrate the advantages of Pixel-Face,
we re-parameterize the 3D Morphable Model (3DMM) into Pixel-3DM using the
collected data. We show that the obtained Pixel-3DM is better in modeling a
wide range of face shapes and expressions. We also carefully benchmark existing
3D face reconstruction methods on our dataset. Moreover, Pixel-Face serves as
an effective training source. We observe that the performance of current face
reconstruction models significantly improves both on existing benchmarks and
Pixel-Face after being fine-tuned using our newly collected data. Extensive
experiments demonstrate the effectiveness of Pixel-3DM and the usefulness of
Pixel-Face.
Related papers
- Pixel3DMM: Versatile Screen-Space Priors for Single-Image 3D Face Reconstruction [46.52887358194364]
We propose Pixel3DMM, a set of highly-generalized vision transformers which predict per-pixel geometric cues.
We train our model by registering three high-quality 3D face datasets against the FLAME mesh topology.
Our method outperforms the most competitive baselines by over 15% in terms of geometric accuracy for posed facial expressions.
arXiv Detail & Related papers (2025-05-01T15:47:03Z) - 3D Human Reconstruction in the Wild with Synthetic Data Using Generative Models [52.96248836582542]
We propose an effective approach based on recent diffusion models, termed HumanWild, which can effortlessly generate human images and corresponding 3D mesh annotations.
By exclusively employing generative models, we generate large-scale in-the-wild human images and high-quality annotations, eliminating the need for real-world data collection.
arXiv Detail & Related papers (2024-03-17T06:31:16Z) - What You See is What You GAN: Rendering Every Pixel for High-Fidelity
Geometry in 3D GANs [82.3936309001633]
3D-aware Generative Adversarial Networks (GANs) have shown remarkable progress in learning to generate multi-view-consistent images and 3D geometries.
Yet, the significant memory and computational costs of dense sampling in volume rendering have forced 3D GANs to adopt patch-based training or employ low-resolution rendering with post-processing 2D super resolution.
We propose techniques to scale neural volume rendering to the much higher resolution of native 2D images, thereby resolving fine-grained 3D geometry with unprecedented detail.
arXiv Detail & Related papers (2024-01-04T18:50:38Z) - A Large-Scale 3D Face Mesh Video Dataset via Neural Re-parameterized
Optimization [17.938604013181426]
We propose NeuFace, a 3D face mesh pseudo annotation method on videos.
We annotate the per-view/frame accurate and consistent face meshes on large-scale face videos, called the NeuFace-dataset.
By exploiting the naturalness and diversity of 3D faces in our dataset, we demonstrate the usefulness of our dataset for 3D face-related tasks.
arXiv Detail & Related papers (2023-10-04T23:24:22Z) - A lightweight 3D dense facial landmark estimation model from position
map data [0.8508775813669867]
We propose a pipeline to create a dense keypoint training dataset containing 520 key points across the whole face.
We train a lightweight MobileNet-based regressor model with the generated data.
Experimental results show that our trained model outperforms many of the existing methods in spite of its lower model size and minimal computational cost.
arXiv Detail & Related papers (2023-08-29T09:53:10Z) - Beyond 3DMM: Learning to Capture High-fidelity 3D Face Shape [77.95154911528365]
3D Morphable Model (3DMM) fitting has widely benefited face analysis due to its strong 3D priori.
Previous reconstructed 3D faces suffer from degraded visual verisimilitude due to the loss of fine-grained geometry.
This paper proposes a complete solution to capture the personalized shape so that the reconstructed shape looks identical to the corresponding person.
arXiv Detail & Related papers (2022-04-09T03:46:18Z) - AvatarMe++: Facial Shape and BRDF Inference with Photorealistic
Rendering-Aware GANs [119.23922747230193]
We introduce the first method that is able to reconstruct render-ready 3D facial geometry and BRDF from a single "in-the-wild" image.
Our method outperforms the existing arts by a significant margin and reconstructs high-resolution 3D faces from a single low-resolution image.
arXiv Detail & Related papers (2021-12-11T11:36:30Z) - FaceScape: 3D Facial Dataset and Benchmark for Single-View 3D Face
Reconstruction [29.920622006999732]
We present a large-scale detailed 3D face dataset, FaceScape, and the corresponding benchmark to evaluate single-view facial 3D reconstruction.
By training on FaceScape data, a novel algorithm is proposed to predict elaborate riggable 3D face models from a single image input.
We also use FaceScape data to generate the in-the-wild and in-the-lab benchmark to evaluate recent methods of single-view face reconstruction.
arXiv Detail & Related papers (2021-11-01T16:48:34Z) - Pix2Surf: Learning Parametric 3D Surface Models of Objects from Images [64.53227129573293]
We investigate the problem of learning to generate 3D parametric surface representations for novel object instances, as seen from one or more views.
We design neural networks capable of generating high-quality parametric 3D surfaces which are consistent between views.
Our method is supervised and trained on a public dataset of shapes from common object categories.
arXiv Detail & Related papers (2020-08-18T06:33:40Z) - FaceScape: a Large-scale High Quality 3D Face Dataset and Detailed
Riggable 3D Face Prediction [39.95272819738226]
We present a novel algorithm that is able to predict elaborate riggable 3D face models from a single image input.
FaceScape dataset provides 18,760 textured 3D faces, captured from 938 subjects and each with 20 specific expressions.
arXiv Detail & Related papers (2020-03-31T07:11:08Z) - AvatarMe: Realistically Renderable 3D Facial Reconstruction
"in-the-wild" [105.28776215113352]
AvatarMe is the first method that is able to reconstruct photorealistic 3D faces from a single "in-the-wild" image with an increasing level of detail.
It outperforms the existing arts by a significant margin and reconstructs authentic, 4K by 6K-resolution 3D faces from a single low-resolution image.
arXiv Detail & Related papers (2020-03-30T22:17:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.