Unsupervised Traffic Scene Generation with Synthetic 3D Scene Graphs
- URL: http://arxiv.org/abs/2303.08473v1
- Date: Wed, 15 Mar 2023 09:26:29 GMT
- Title: Unsupervised Traffic Scene Generation with Synthetic 3D Scene Graphs
- Authors: Artem Savkin, Rachid Ellouze, Nassir Navab, Federico Tombari
- Abstract summary: We propose a method based on domain-invariant scene representation to directly synthesize traffic scene imagery without rendering.
Specifically, we rely on synthetic scene graphs as our internal representation and introduce an unsupervised neural network architecture for realistic traffic scene synthesis.
- Score: 83.9783063609389
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Image synthesis driven by computer graphics achieved recently a remarkable
realism, yet synthetic image data generated this way reveals a significant
domain gap with respect to real-world data. This is especially true in
autonomous driving scenarios, which represent a critical aspect for overcoming
utilizing synthetic data for training neural networks. We propose a method
based on domain-invariant scene representation to directly synthesize traffic
scene imagery without rendering. Specifically, we rely on synthetic scene
graphs as our internal representation and introduce an unsupervised neural
network architecture for realistic traffic scene synthesis. We enhance
synthetic scene graphs with spatial information about the scene and demonstrate
the effectiveness of our approach through scene manipulation.
Related papers
- Synthesizing Traffic Datasets using Graph Neural Networks [2.444217495283211]
This paper introduces a novel methodology for bridging this sim-real' gap by creating photorealistic images from 2D traffic simulations and recorded junction footage.
We propose a novel image generation approach, integrating a Conditional Generative Adversarial Network with a Graph Neural Network (GNN) to facilitate the creation of realistic urban traffic images.
arXiv Detail & Related papers (2023-12-08T13:24:19Z) - Neural Radiance Transfer Fields for Relightable Novel-view Synthesis
with Global Illumination [63.992213016011235]
We propose a method for scene relighting under novel views by learning a neural precomputed radiance transfer function.
Our method can be solely supervised on a set of real images of the scene under a single unknown lighting condition.
Results show that the recovered disentanglement of scene parameters improves significantly over the current state of the art.
arXiv Detail & Related papers (2022-07-27T16:07:48Z) - Leveraging Deepfakes to Close the Domain Gap between Real and Synthetic
Images in Facial Capture Pipelines [8.366597450893456]
We propose an end-to-end pipeline for building and tracking 3D facial models from personalized in-the-wild video data.
We present a method for automatic data curation and retrieval based on a hierarchical clustering framework typical of collision algorithms in traditional computer graphics pipelines.
We outline how we train a motion capture regressor, leveraging the aforementioned techniques to avoid the need for real-world ground truth data.
arXiv Detail & Related papers (2022-04-22T15:09:49Z) - Towards 3D Scene Understanding by Referring Synthetic Models [65.74211112607315]
Methods typically alleviate on-extensive annotations on real scene scans.
We explore how synthetic models rely on real scene categories of synthetic features to a unified feature space.
Experiments show that our method achieves the average mAP of 46.08% on the ScanNet S3DIS dataset and 55.49% by learning datasets.
arXiv Detail & Related papers (2022-03-20T13:06:15Z) - SceneGen: Learning to Generate Realistic Traffic Scenes [92.98412203941912]
We present SceneGen, a neural autoregressive model of traffic scenes that eschews the need for rules and distributions.
We demonstrate SceneGen's ability to faithfully model distributions of real traffic scenes.
arXiv Detail & Related papers (2021-01-16T22:51:43Z) - Neural Scene Graphs for Dynamic Scenes [57.65413768984925]
We present the first neural rendering method that decomposes dynamic scenes into scene graphs.
We learn implicitly encoded scenes, combined with a jointly learned latent representation to describe objects with a single implicit function.
arXiv Detail & Related papers (2020-11-20T12:37:10Z) - Intrinsic Autoencoders for Joint Neural Rendering and Intrinsic Image
Decomposition [67.9464567157846]
We propose an autoencoder for joint generation of realistic images from synthetic 3D models while simultaneously decomposing real images into their intrinsic shape and appearance properties.
Our experiments confirm that a joint treatment of rendering and decomposition is indeed beneficial and that our approach outperforms state-of-the-art image-to-image translation baselines both qualitatively and quantitatively.
arXiv Detail & Related papers (2020-06-29T12:53:58Z) - From Real to Synthetic and Back: Synthesizing Training Data for
Multi-Person Scene Understanding [0.7519872646378835]
We present a method for synthesizing naturally looking images of multiple people interacting in a specific scenario.
These images benefit from the advantages of synthetic data: being fully controllable and fully annotated with any type of standard or custom-defined ground truth.
To reduce the synthetic-to-real domain gap, we introduce a pipeline consisting of the following steps.
arXiv Detail & Related papers (2020-06-03T09:02:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.