Freestyle 3D-Aware Portrait Synthesis Based on Compositional Generative
Priors
- URL: http://arxiv.org/abs/2306.15419v3
- Date: Sun, 24 Dec 2023 08:35:38 GMT
- Title: Freestyle 3D-Aware Portrait Synthesis Based on Compositional Generative
Priors
- Authors: Tianxiang Ma, Kang Zhao, Jianxin Sun, Yingya Zhang, Jing Dong
- Abstract summary: We propose a novel text-driven 3D-aware portrait synthesis framework.
Specifically, for a given portrait style prompt, we first composite two generative priors, a 3D-aware GAN generator and a text-guided image editor.
Then we map the special style domain of this set to our proposed 3D latent feature generator and obtain a 3D representation containing the given style information.
- Score: 12.663585627797863
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Efficiently generating a freestyle 3D portrait with high quality and
3D-consistency is a promising yet challenging task. The portrait styles
generated by most existing methods are usually restricted by their 3D
generators, which are learned in specific facial datasets, such as FFHQ. To get
the diverse 3D portraits, one can build a large-scale multi-style database to
retrain a 3D-aware generator, or use a off-the-shelf tool to do the style
translation. However, the former is time-consuming due to data collection and
training process, the latter may destroy the multi-view consistency. To tackle
this problem, we propose a novel text-driven 3D-aware portrait synthesis
framework that can generate out-of-distribution portrait styles. Specifically,
for a given portrait style prompt, we first composite two generative priors, a
3D-aware GAN generator and a text-guided image editor, to quickly construct a
few-shot stylized portrait set. Then we map the special style domain of this
set to our proposed 3D latent feature generator and obtain a 3D representation
containing the given style information. Finally we use a pre-trained 3D
renderer to generate view-consistent stylized portraits from the 3D
representation. Extensive experimental results show that our method is capable
of synthesizing high-quality 3D portraits with specified styles in a few
minutes, outperforming the state-of-the-art.
Related papers
- Portrait3D: Text-Guided High-Quality 3D Portrait Generation Using Pyramid Representation and GANs Prior [21.18119394249007]
We present Portrait3D, a novel neural rendering-based framework with a novel joint geometry-appearance prior to achieve text-to-3D-portrait generation.
Our experimental results show that Portrait3D can produce realistic, high-quality, and canonical 3D portraits that align with the prompt.
arXiv Detail & Related papers (2024-04-16T08:52:42Z) - RealmDreamer: Text-Driven 3D Scene Generation with Inpainting and Depth Diffusion [39.03289977892935]
RealmDreamer is a technique for generation of general forward-facing 3D scenes from text descriptions.
Our technique does not require video or multi-view data and can synthesize a variety of high-quality 3D scenes in different styles.
arXiv Detail & Related papers (2024-04-10T17:57:41Z) - Sketch3D: Style-Consistent Guidance for Sketch-to-3D Generation [55.73399465968594]
This paper proposes a novel generation paradigm Sketch3D to generate realistic 3D assets with shape aligned with the input sketch and color matching the textual description.
Three strategies are designed to optimize 3D Gaussians, i.e., structural optimization via a distribution transfer mechanism, color optimization with a straightforward MSE loss and sketch similarity optimization with a CLIP-based geometric similarity loss.
arXiv Detail & Related papers (2024-04-02T11:03:24Z) - HyperStyle3D: Text-Guided 3D Portrait Stylization via Hypernetworks [101.36230756743106]
This paper is inspired by the success of 3D-aware GANs that bridge 2D and 3D domains with 3D fields as the intermediate representation for rendering 2D images.
We propose a novel method, dubbed HyperStyle3D, based on 3D-aware GANs for 3D portrait stylization.
arXiv Detail & Related papers (2023-04-19T07:22:05Z) - AgileGAN3D: Few-Shot 3D Portrait Stylization by Augmented Transfer
Learning [80.67196184480754]
We propose a novel framework emphAgileGAN3D that can produce 3D artistically appealing portraits with detailed geometry.
New stylization can be obtained with just a few (around 20) unpaired 2D exemplars.
Our pipeline demonstrates strong capability in turning user photos into a diverse range of 3D artistic portraits.
arXiv Detail & Related papers (2023-03-24T23:04:20Z) - CC3D: Layout-Conditioned Generation of Compositional 3D Scenes [49.281006972028194]
We introduce CC3D, a conditional generative model that synthesizes complex 3D scenes conditioned on 2D semantic scene layouts.
Our evaluations on synthetic 3D-FRONT and real-world KITTI-360 datasets demonstrate that our model generates scenes of improved visual and geometric quality.
arXiv Detail & Related papers (2023-03-21T17:59:02Z) - Make Your Brief Stroke Real and Stereoscopic: 3D-Aware Simplified Sketch
to Portrait Generation [51.64832538714455]
Existing studies only generate portraits in the 2D plane with fixed views, making the results less vivid.
In this paper, we present Stereoscopic Simplified Sketch-to-Portrait (SSSP), which explores the possibility of creating Stereoscopic 3D-aware portraits.
Our key insight is to design sketch-aware constraints that can fully exploit the prior knowledge of a tri-plane-based 3D-aware generative model.
arXiv Detail & Related papers (2023-02-14T06:28:42Z) - PV3D: A 3D Generative Model for Portrait Video Generation [94.96025739097922]
We propose PV3D, the first generative framework that can synthesize multi-view consistent portrait videos.
PV3D is able to support many downstream applications such as animating static portraits and view-consistent video motion editing.
arXiv Detail & Related papers (2022-12-13T05:42:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.