Get3DHuman: Lifting StyleGAN-Human into a 3D Generative Model using
Pixel-aligned Reconstruction Priors
- URL: http://arxiv.org/abs/2302.01162v5
- Date: Mon, 24 Jul 2023 09:41:07 GMT
- Title: Get3DHuman: Lifting StyleGAN-Human into a 3D Generative Model using
Pixel-aligned Reconstruction Priors
- Authors: Zhangyang Xiong, Di Kang, Derong Jin, Weikai Chen, Linchao Bao,
Shuguang Cui, Xiaoguang Han
- Abstract summary: Get3DHuman is a novel 3D human framework that can significantly boost the realism and diversity of the generated outcomes.
Our key observation is that the 3D generator can profit from human-related priors learned through 2D human generators and 3D reconstructors.
- Score: 56.192682114114724
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Fast generation of high-quality 3D digital humans is important to a vast
number of applications ranging from entertainment to professional concerns.
Recent advances in differentiable rendering have enabled the training of 3D
generative models without requiring 3D ground truths. However, the quality of
the generated 3D humans still has much room to improve in terms of both
fidelity and diversity. In this paper, we present Get3DHuman, a novel 3D human
framework that can significantly boost the realism and diversity of the
generated outcomes by only using a limited budget of 3D ground-truth data. Our
key observation is that the 3D generator can profit from human-related priors
learned through 2D human generators and 3D reconstructors. Specifically, we
bridge the latent space of Get3DHuman with that of StyleGAN-Human via a
specially-designed prior network, where the input latent code is mapped to the
shape and texture feature volumes spanned by the pixel-aligned 3D
reconstructor. The outcomes of the prior network are then leveraged as the
supervisory signals for the main generator network. To ensure effective
training, we further propose three tailored losses applied to the generated
feature volumes and the intermediate feature maps. Extensive experiments
demonstrate that Get3DHuman greatly outperforms the other state-of-the-art
approaches and can support a wide range of applications including shape
interpolation, shape re-texturing, and single-view reconstruction through
latent inversion.
Related papers
- DiffTF++: 3D-aware Diffusion Transformer for Large-Vocabulary 3D Generation [53.20147419879056]
We introduce a diffusion-based feed-forward framework to address challenges with a single model.
Building upon our 3D-aware Diffusion model with TransFormer, we propose a stronger version for 3D generation, i.e., DiffTF++.
Experiments on ShapeNet and OmniObject3D convincingly demonstrate the effectiveness of our proposed modules.
arXiv Detail & Related papers (2024-05-13T17:59:51Z) - IM-3D: Iterative Multiview Diffusion and Reconstruction for High-Quality
3D Generation [96.32684334038278]
In this paper, we explore the design space of text-to-3D models.
We significantly improve multi-view generation by considering video instead of image generators.
Our new method, IM-3D, reduces the number of evaluations of the 2D generator network 10-100x.
arXiv Detail & Related papers (2024-02-13T18:59:51Z) - Progress and Prospects in 3D Generative AI: A Technical Overview
including 3D human [51.58094069317723]
This paper aims to provide a comprehensive overview and summary of the relevant papers published mostly during the latter half year of 2023.
It will begin by discussing the AI generated object models in 3D, followed by the generated 3D human models, and finally, the generated 3D human motions, culminating in a conclusive summary and a vision for the future.
arXiv Detail & Related papers (2024-01-05T03:41:38Z) - HumanLiff: Layer-wise 3D Human Generation with Diffusion Model [55.891036415316876]
Existing 3D human generative models mainly generate a clothed 3D human as an undetectable 3D model in a single pass.
We propose HumanLiff, the first layer-wise 3D human generative model with a unified diffusion process.
arXiv Detail & Related papers (2023-08-18T17:59:04Z) - 3D GANs and Latent Space: A comprehensive survey [0.0]
3D GANs are a new type of generative model used for 3D reconstruction, point cloud reconstruction, and 3D semantic scene completion.
The choice of distribution for noise is critical as it represents the latent space.
In this work, we explore the latent space and 3D GANs, examine several GAN variants and training methods to gain insights into improving 3D GAN training, and suggest potential future directions for further research.
arXiv Detail & Related papers (2023-04-08T06:36:07Z) - HumanGen: Generating Human Radiance Fields with Explicit Priors [19.5166920467636]
HumanGen is a novel 3D human generation scheme with detailed geometry and realistic free-view rendering.
It explicitly marries the 3D human generation with various priors from the 2D generator and 3D reconstructor of humans through the design of "anchor image"
arXiv Detail & Related papers (2022-12-10T15:27:48Z) - GET3D: A Generative Model of High Quality 3D Textured Shapes Learned
from Images [72.15855070133425]
We introduce GET3D, a Generative model that directly generates Explicit Textured 3D meshes with complex topology, rich geometric details, and high-fidelity textures.
GET3D is able to generate high-quality 3D textured meshes, ranging from cars, chairs, animals, motorbikes and human characters to buildings.
arXiv Detail & Related papers (2022-09-22T17:16:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.