SceneGen: Learning to Generate Realistic Traffic Scenes
- URL: http://arxiv.org/abs/2101.06541v1
- Date: Sat, 16 Jan 2021 22:51:43 GMT
- Title: SceneGen: Learning to Generate Realistic Traffic Scenes
- Authors: Shuhan Tan, Kelvin Wong, Shenlong Wang, Sivabalan Manivasagam, Mengye
Ren, Raquel Urtasun
- Abstract summary: We present SceneGen, a neural autoregressive model of traffic scenes that eschews the need for rules and distributions.
We demonstrate SceneGen's ability to faithfully model distributions of real traffic scenes.
- Score: 92.98412203941912
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We consider the problem of generating realistic traffic scenes automatically.
Existing methods typically insert actors into the scene according to a set of
hand-crafted heuristics and are limited in their ability to model the true
complexity and diversity of real traffic scenes, thus inducing a content gap
between synthesized traffic scenes versus real ones. As a result, existing
simulators lack the fidelity necessary to train and test self-driving vehicles.
To address this limitation, we present SceneGen, a neural autoregressive model
of traffic scenes that eschews the need for rules and heuristics. In
particular, given the ego-vehicle state and a high definition map of
surrounding area, SceneGen inserts actors of various classes into the scene and
synthesizes their sizes, orientations, and velocities. We demonstrate on two
large-scale datasets SceneGen's ability to faithfully model distributions of
real traffic scenes. Moreover, we show that SceneGen coupled with sensor
simulation can be used to train perception models that generalize to the real
world.
Related papers
- AutoSplat: Constrained Gaussian Splatting for Autonomous Driving Scene Reconstruction [17.600027937450342]
AutoSplat is a framework employing Gaussian splatting to achieve highly realistic reconstructions of autonomous driving scenes.
Our method enables multi-view consistent simulation of challenging scenarios including lane changes.
arXiv Detail & Related papers (2024-07-02T18:36:50Z) - DragTraffic: Interactive and Controllable Traffic Scene Generation for Autonomous Driving [10.90477019946728]
DragTraffic is a general, interactive, and controllable traffic scene generation framework based on conditional diffusion.
We employ a regression model to provide a general initial solution and a refinement process based on the conditional diffusion model to ensure diversity.
Experiments on a real-world driving dataset show that DragTraffic outperforms existing methods in terms of authenticity, diversity, and freedom.
arXiv Detail & Related papers (2024-04-19T04:49:28Z) - Language Conditioned Traffic Generation [37.71751991840586]
LCTGen is a large language model with a transformer-based decoder architecture that selects likely map locations from a dataset of maps.
It produces an initial traffic distribution, as well as the dynamics of each vehicle.
LCTGen outperforms prior work in both unconditional and conditional traffic scene generation in terms of realism and fidelity.
arXiv Detail & Related papers (2023-07-16T05:10:32Z) - Unsupervised Traffic Scene Generation with Synthetic 3D Scene Graphs [83.9783063609389]
We propose a method based on domain-invariant scene representation to directly synthesize traffic scene imagery without rendering.
Specifically, we rely on synthetic scene graphs as our internal representation and introduce an unsupervised neural network architecture for realistic traffic scene synthesis.
arXiv Detail & Related papers (2023-03-15T09:26:29Z) - TrafficBots: Towards World Models for Autonomous Driving Simulation and
Motion Prediction [149.5716746789134]
We show data-driven traffic simulation can be formulated as a world model.
We present TrafficBots, a multi-agent policy built upon motion prediction and end-to-end driving.
Experiments on the open motion dataset show TrafficBots can simulate realistic multi-agent behaviors.
arXiv Detail & Related papers (2023-03-07T18:28:41Z) - Street-View Image Generation from a Bird's-Eye View Layout [95.36869800896335]
Bird's-Eye View (BEV) Perception has received increasing attention in recent years.
Data-driven simulation for autonomous driving has been a focal point of recent research.
We propose BEVGen, a conditional generative model that synthesizes realistic and spatially consistent surrounding images.
arXiv Detail & Related papers (2023-01-11T18:39:34Z) - READ: Large-Scale Neural Scene Rendering for Autonomous Driving [21.144110676687667]
A large-scale neural rendering method is proposed to synthesize the autonomous driving scene.
Our model can not only synthesize realistic driving scenes but also stitch and edit driving scenes.
arXiv Detail & Related papers (2022-05-11T14:02:14Z) - DriveGAN: Towards a Controllable High-Quality Neural Simulation [147.6822288981004]
We introduce a novel high-quality neural simulator referred to as DriveGAN.
DriveGAN achieves controllability by disentangling different components without supervision.
We train DriveGAN on multiple datasets, including 160 hours of real-world driving data.
arXiv Detail & Related papers (2021-04-30T15:30:05Z) - TrafficSim: Learning to Simulate Realistic Multi-Agent Behaviors [74.67698916175614]
We propose TrafficSim, a multi-agent behavior model for realistic traffic simulation.
In particular, we leverage an implicit latent variable model to parameterize a joint actor policy.
We show TrafficSim generates significantly more realistic and diverse traffic scenarios as compared to a diverse set of baselines.
arXiv Detail & Related papers (2021-01-17T00:29:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.