Normalized Avatar Synthesis Using StyleGAN and Perceptual Refinement
- URL: http://arxiv.org/abs/2106.11423v1
- Date: Mon, 21 Jun 2021 21:57:16 GMT
- Title: Normalized Avatar Synthesis Using StyleGAN and Perceptual Refinement
- Authors: Huiwen Luo, Koki Nagano, Han-Wei Kung, Mclean Goldwhite, Qingguo Xu,
Zejian Wang, Lingyu Wei, Liwen Hu, Hao Li
- Abstract summary: We introduce a highly robust GAN-based framework for digitizing a normalized 3D avatar of a person from a single unconstrained photo.
While the input image can be of a smiling person or taken in extreme lighting conditions, our method can reliably produce a high-quality textured model of a person's face.
- Score: 11.422683083130577
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: We introduce a highly robust GAN-based framework for digitizing a normalized
3D avatar of a person from a single unconstrained photo. While the input image
can be of a smiling person or taken in extreme lighting conditions, our method
can reliably produce a high-quality textured model of a person's face in
neutral expression and skin textures under diffuse lighting condition.
Cutting-edge 3D face reconstruction methods use non-linear morphable face
models combined with GAN-based decoders to capture the likeness and details of
a person but fail to produce neutral head models with unshaded albedo textures
which is critical for creating relightable and animation-friendly avatars for
integration in virtual environments. The key challenges for existing methods to
work is the lack of training and ground truth data containing normalized 3D
faces. We propose a two-stage approach to address this problem. First, we adopt
a highly robust normalized 3D face generator by embedding a non-linear
morphable face model into a StyleGAN2 network. This allows us to generate
detailed but normalized facial assets. This inference is then followed by a
perceptual refinement step that uses the generated assets as regularization to
cope with the limited available training samples of normalized faces. We
further introduce a Normalized Face Dataset, which consists of a combination
photogrammetry scans, carefully selected photographs, and generated fake people
with neutral expressions in diffuse lighting conditions. While our prepared
dataset contains two orders of magnitude less subjects than cutting edge
GAN-based 3D facial reconstruction methods, we show that it is possible to
produce high-quality normalized face models for very challenging unconstrained
input images, and demonstrate superior performance to the current
state-of-the-art.
Related papers
- Single Image, Any Face: Generalisable 3D Face Generation [59.9369171926757]
We propose a novel model, Gen3D-Face, which generates 3D human faces with unconstrained single image input.
To the best of our knowledge, this is the first attempt and benchmark for creating photorealistic 3D human face avatars from single images.
arXiv Detail & Related papers (2024-09-25T14:56:37Z) - Single-Shot Implicit Morphable Faces with Consistent Texture
Parameterization [91.52882218901627]
We propose a novel method for constructing implicit 3D morphable face models that are both generalizable and intuitive for editing.
Our method improves upon photo-realism, geometry, and expression accuracy compared to state-of-the-art methods.
arXiv Detail & Related papers (2023-05-04T17:58:40Z) - 3DMM-RF: Convolutional Radiance Fields for 3D Face Modeling [111.98096975078158]
We introduce a style-based generative network that synthesizes in one pass all and only the required rendering samples of a neural radiance field.
We show that this model can accurately be fit to "in-the-wild" facial images of arbitrary pose and illumination, extract the facial characteristics, and be used to re-render the face in controllable conditions.
arXiv Detail & Related papers (2022-09-15T15:28:45Z) - AvatarMe++: Facial Shape and BRDF Inference with Photorealistic
Rendering-Aware GANs [119.23922747230193]
We introduce the first method that is able to reconstruct render-ready 3D facial geometry and BRDF from a single "in-the-wild" image.
Our method outperforms the existing arts by a significant margin and reconstructs high-resolution 3D faces from a single low-resolution image.
arXiv Detail & Related papers (2021-12-11T11:36:30Z) - MOST-GAN: 3D Morphable StyleGAN for Disentangled Face Image Manipulation [69.35523133292389]
We propose a framework that a priori models physical attributes of the face explicitly, thus providing disentanglement by design.
Our method, MOST-GAN, integrates the expressive power and photorealism of style-based GANs with the physical disentanglement and flexibility of nonlinear 3D morphable models.
It achieves photorealistic manipulation of portrait images with fully disentangled 3D control over their physical attributes, enabling extreme manipulation of lighting, facial expression, and pose variations up to full profile view.
arXiv Detail & Related papers (2021-11-01T15:53:36Z) - OSTeC: One-Shot Texture Completion [86.23018402732748]
We propose an unsupervised approach for one-shot 3D facial texture completion.
The proposed approach rotates an input image in 3D and fill-in the unseen regions by reconstructing the rotated image in a 2D face generator.
We frontalize the target image by projecting the completed texture into the generator.
arXiv Detail & Related papers (2020-12-30T23:53:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.