Deep CG2Real: Synthetic-to-Real Translation via Image Disentanglement
- URL: http://arxiv.org/abs/2003.12649v1
- Date: Fri, 27 Mar 2020 21:45:41 GMT
- Title: Deep CG2Real: Synthetic-to-Real Translation via Image Disentanglement
- Authors: Sai Bi, Kalyan Sunkavalli, Federico Perazzi, Eli Shechtman, Vladimir
Kim, Ravi Ramamoorthi
- Abstract summary: Training an unpaired synthetic-to-real translation network in image space is severely under-constrained.
We propose a semi-supervised approach that operates on the disentangled shading and albedo layers of the image.
Our two-stage pipeline first learns to predict accurate shading in a supervised fashion using physically-based renderings as targets.
- Score: 78.58603635621591
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present a method to improve the visual realism of low-quality, synthetic
images, e.g. OpenGL renderings. Training an unpaired synthetic-to-real
translation network in image space is severely under-constrained and produces
visible artifacts. Instead, we propose a semi-supervised approach that operates
on the disentangled shading and albedo layers of the image. Our two-stage
pipeline first learns to predict accurate shading in a supervised fashion using
physically-based renderings as targets, and further increases the realism of
the textures and shading with an improved CycleGAN network. Extensive
evaluations on the SUNCG indoor scene dataset demonstrate that our approach
yields more realistic images compared to other state-of-the-art approaches.
Furthermore, networks trained on our generated "real" images predict more
accurate depth and normals than domain adaptation approaches, suggesting that
improving the visual realism of the images can be more effective than imposing
task-specific losses.
Related papers
- Deep Domain Adaptation: A Sim2Real Neural Approach for Improving Eye-Tracking Systems [80.62854148838359]
Eye image segmentation is a critical step in eye tracking that has great influence over the final gaze estimate.
We use dimensionality-reduction techniques to measure the overlap between the target eye images and synthetic training data.
Our methods result in robust, improved performance when tackling the discrepancy between simulation and real-world data samples.
arXiv Detail & Related papers (2024-03-23T22:32:06Z) - Photo-realistic Neural Domain Randomization [37.42597274391271]
We show that the recent progress in neural rendering enables a new unified approach we call Photo-realistic Neural Domain Randomization (PNDR)
Our approach is modular, composed of different neural networks for materials, lighting, and rendering, thus enabling randomization of different key image generation components in a differentiable pipeline.
Our experiments show that training with PNDR enables generalization to novel scenes and significantly outperforms the state of the art in terms of real-world transfer.
arXiv Detail & Related papers (2022-10-23T09:45:27Z) - Enhancing Photorealism Enhancement [83.88433283714461]
We present an approach to enhancing the realism of synthetic images using a convolutional network.
We analyze scene layout distributions in commonly used datasets and find that they differ in important ways.
We report substantial gains in stability and realism in comparison to recent image-to-image translation methods.
arXiv Detail & Related papers (2021-05-10T19:00:49Z) - Inverting Generative Adversarial Renderer for Face Reconstruction [58.45125455811038]
In this work, we introduce a novel Generative Adversa Renderer (GAR)
GAR learns to model the complicated real-world image, instead of relying on the graphics rules, it is capable of producing realistic images.
Our method achieves state-of-the-art performances on multiple face reconstruction.
arXiv Detail & Related papers (2021-05-06T04:16:06Z) - Intrinsic Autoencoders for Joint Neural Rendering and Intrinsic Image
Decomposition [67.9464567157846]
We propose an autoencoder for joint generation of realistic images from synthetic 3D models while simultaneously decomposing real images into their intrinsic shape and appearance properties.
Our experiments confirm that a joint treatment of rendering and decomposition is indeed beneficial and that our approach outperforms state-of-the-art image-to-image translation baselines both qualitatively and quantitatively.
arXiv Detail & Related papers (2020-06-29T12:53:58Z) - Two-shot Spatially-varying BRDF and Shape Estimation [89.29020624201708]
We propose a novel deep learning architecture with a stage-wise estimation of shape and SVBRDF.
We create a large-scale synthetic training dataset with domain-randomized geometry and realistic materials.
Experiments on both synthetic and real-world datasets show that our network trained on a synthetic dataset can generalize well to real-world images.
arXiv Detail & Related papers (2020-04-01T12:56:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.