Realistic Image Synthesis with Configurable 3D Scene Layouts
- URL: http://arxiv.org/abs/2108.10031v2
- Date: Tue, 24 Aug 2021 05:52:59 GMT
- Title: Realistic Image Synthesis with Configurable 3D Scene Layouts
- Authors: Jaebong Jeong, Janghun Jo, Jingdong Wang, Sunghyun Cho, Jaesik Park
- Abstract summary: We propose a novel approach to realistic-looking image synthesis based on a 3D scene layout.
Our approach takes a 3D scene with semantic class labels as input and trains a 3D scene painting network.
With the trained painting network, realistic-looking images for the input 3D scene can be rendered and manipulated.
- Score: 59.872657806747576
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Recent conditional image synthesis approaches provide high-quality
synthesized images. However, it is still challenging to accurately adjust image
contents such as the positions and orientations of objects, and synthesized
images often have geometrically invalid contents. To provide users with rich
controllability on synthesized images in the aspect of 3D geometry, we propose
a novel approach to realistic-looking image synthesis based on a configurable
3D scene layout. Our approach takes a 3D scene with semantic class labels as
input and trains a 3D scene painting network that synthesizes color values for
the input 3D scene. With the trained painting network, realistic-looking images
for the input 3D scene can be rendered and manipulated. To train the painting
network without 3D color supervision, we exploit an off-the-shelf 2D semantic
image synthesis method. In experiments, we show that our approach produces
images with geometrically correct structures and supports geometric
manipulation such as the change of the viewpoint and object poses as well as
manipulation of the painting style.
Related papers
- 3D-SSGAN: Lifting 2D Semantics for 3D-Aware Compositional Portrait
Synthesis [11.457566989721078]
Existing 3D-aware portrait synthesis methods can generate impressive high-quality images while preserving strong 3D consistency.
We propose 3D-SSGAN, a novel framework for 3D-aware compositional portrait image synthesis.
arXiv Detail & Related papers (2024-01-08T09:41:07Z) - CC3D: Layout-Conditioned Generation of Compositional 3D Scenes [49.281006972028194]
We introduce CC3D, a conditional generative model that synthesizes complex 3D scenes conditioned on 2D semantic scene layouts.
Our evaluations on synthetic 3D-FRONT and real-world KITTI-360 datasets demonstrate that our model generates scenes of improved visual and geometric quality.
arXiv Detail & Related papers (2023-03-21T17:59:02Z) - CompNVS: Novel View Synthesis with Scene Completion [83.19663671794596]
We propose a generative pipeline performing on a sparse grid-based neural scene representation to complete unobserved scene parts.
We process encoded image features in 3D space with a geometry completion network and a subsequent texture inpainting network to extrapolate the missing area.
Photorealistic image sequences can be finally obtained via consistency-relevant differentiable rendering.
arXiv Detail & Related papers (2022-07-23T09:03:13Z) - Style Agnostic 3D Reconstruction via Adversarial Style Transfer [23.304453155586312]
Reconstructing the 3D geometry of an object from an image is a major challenge in computer vision.
We propose an approach that enables a differentiable-based learning of 3D objects from images with backgrounds.
arXiv Detail & Related papers (2021-10-20T21:24:44Z) - Weakly Supervised Learning of Multi-Object 3D Scene Decompositions Using
Deep Shape Priors [69.02332607843569]
PriSMONet is a novel approach for learning Multi-Object 3D scene decomposition and representations from single images.
A recurrent encoder regresses a latent representation of 3D shape, pose and texture of each object from an input RGB image.
We evaluate the accuracy of our model in inferring 3D scene layout, demonstrate its generative capabilities, assess its generalization to real images, and point out benefits of the learned representation.
arXiv Detail & Related papers (2020-10-08T14:49:23Z) - Intrinsic Autoencoders for Joint Neural Rendering and Intrinsic Image
Decomposition [67.9464567157846]
We propose an autoencoder for joint generation of realistic images from synthetic 3D models while simultaneously decomposing real images into their intrinsic shape and appearance properties.
Our experiments confirm that a joint treatment of rendering and decomposition is indeed beneficial and that our approach outperforms state-of-the-art image-to-image translation baselines both qualitatively and quantitatively.
arXiv Detail & Related papers (2020-06-29T12:53:58Z) - 3D Photography using Context-aware Layered Depth Inpainting [50.66235795163143]
We propose a method for converting a single RGB-D input image into a 3D photo.
A learning-based inpainting model synthesizes new local color-and-depth content into the occluded region.
The resulting 3D photos can be efficiently rendered with motion parallax.
arXiv Detail & Related papers (2020-04-09T17:59:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.