Embedding Synthetic Off-Policy Experience for Autonomous Driving via
Zero-Shot Curricula
- URL: http://arxiv.org/abs/2212.01375v1
- Date: Fri, 2 Dec 2022 18:57:21 GMT
- Title: Embedding Synthetic Off-Policy Experience for Autonomous Driving via
Zero-Shot Curricula
- Authors: Eli Bronstein, Sirish Srinivasan, Supratik Paul, Aman Sinha, Matthew
O'Kelly, Payam Nikdel, Shimon Whiteson
- Abstract summary: We show that an agent trained using only a 10% subset of the data performs just as well as an agent trained on the entire dataset.
We then demonstrate that this difficulty score can be used in a zero-shot transfer to generate curricula for an imitation-learning based planning agent.
- Score: 48.58973705935691
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: ML-based motion planning is a promising approach to produce agents that
exhibit complex behaviors, and automatically adapt to novel environments. In
the context of autonomous driving, it is common to treat all available training
data equally. However, this approach produces agents that do not perform
robustly in safety-critical settings, an issue that cannot be addressed by
simply adding more data to the training set - we show that an agent trained
using only a 10% subset of the data performs just as well as an agent trained
on the entire dataset. We present a method to predict the inherent difficulty
of a driving situation given data collected from a fleet of autonomous vehicles
deployed on public roads. We then demonstrate that this difficulty score can be
used in a zero-shot transfer to generate curricula for an imitation-learning
based planning agent. Compared to training on the entire unbiased training
dataset, we show that prioritizing difficult driving scenarios both reduces
collisions by 15% and increases route adherence by 14% in closed-loop
evaluation, all while using only 10% of the training data.
Related papers
- Learning to Drive via Asymmetric Self-Play [34.56873945538085]
We propose asymmetric self-play to scale beyond real data with challenging, solvable, and realistic synthetic scenarios.
Our approach pairs a teacher that learns to generate scenarios it can solve but the student cannot, with a student that learns to solve them.
Our policies further zero-shot transfer to generate training data for end-to-end autonomy, significantly outperforming state-of-the-art adversarial approaches.
arXiv Detail & Related papers (2024-09-26T18:55:38Z) - Data-efficient Trajectory Prediction via Coreset Selection [4.682090083225856]
Training trajectory prediction models is challenging in two ways.
Easy-medium driving scenarios often overwhelmingly dominate the dataset.
We propose a novel data-efficient training method based on coreset selection.
arXiv Detail & Related papers (2024-09-25T22:00:11Z) - Mitigating Covariate Shift in Imitation Learning for Autonomous Vehicles Using Latent Space Generative World Models [60.87795376541144]
A world model is a neural network capable of predicting an agent's next state given past states and actions.
During end-to-end training, our policy learns how to recover from errors by aligning with states observed in human demonstrations.
We present qualitative and quantitative results, demonstrating significant improvements upon prior state of the art in closed-loop testing.
arXiv Detail & Related papers (2024-09-25T06:48:25Z) - Autonomous Vehicle Controllers From End-to-End Differentiable Simulation [60.05963742334746]
We propose a differentiable simulator and design an analytic policy gradients (APG) approach to training AV controllers.
Our proposed framework brings the differentiable simulator into an end-to-end training loop, where gradients of environment dynamics serve as a useful prior to help the agent learn a more grounded policy.
We find significant improvements in performance and robustness to noise in the dynamics, as well as overall more intuitive human-like handling.
arXiv Detail & Related papers (2024-09-12T11:50:06Z) - Perception Without Vision for Trajectory Prediction: Ego Vehicle Dynamics as Scene Representation for Efficient Active Learning in Autonomous Driving [0.0]
We propose methods for clustering trajectory-states and sampling strategies in an active learning framework.
By integrating trajectory-state-informed active learning, we demonstrate that more efficient and robust autonomous driving systems are possible.
arXiv Detail & Related papers (2024-05-15T02:54:11Z) - Pre-training on Synthetic Driving Data for Trajectory Prediction [61.520225216107306]
We propose a pipeline-level solution to mitigate the issue of data scarcity in trajectory forecasting.
We adopt HD map augmentation and trajectory synthesis for generating driving data, and then we learn representations by pre-training on them.
We conduct extensive experiments to demonstrate the effectiveness of our data expansion and pre-training strategies.
arXiv Detail & Related papers (2023-09-18T19:49:22Z) - Rethinking Closed-loop Training for Autonomous Driving [82.61418945804544]
We present the first empirical study which analyzes the effects of different training benchmark designs on the success of learning agents.
We propose trajectory value learning (TRAVL), an RL-based driving agent that performs planning with multistep look-ahead.
Our experiments show that TRAVL can learn much faster and produce safer maneuvers compared to all the baselines.
arXiv Detail & Related papers (2023-06-27T17:58:39Z) - Learning Interactive Driving Policies via Data-driven Simulation [125.97811179463542]
Data-driven simulators promise high data-efficiency for driving policy learning.
Small underlying datasets often lack interesting and challenging edge cases for learning interactive driving.
We propose a simulation method that uses in-painted ado vehicles for learning robust driving policies.
arXiv Detail & Related papers (2021-11-23T20:14:02Z) - Equilibrium Inverse Reinforcement Learning for Ride-hailing Vehicle
Network [1.599072005190786]
We formulate the problem of passenger-vehicle matching in a sparsely connected graph.
We propose an algorithm to derive an equilibrium policy in a multi-agent environment.
arXiv Detail & Related papers (2021-02-13T03:18:44Z) - Meta Reinforcement Learning-Based Lane Change Strategy for Autonomous
Vehicles [11.180588185127892]
Supervised learning algorithms can generalize to new environments by training on a large amount of labeled data.
It can be often impractical or cost-prohibitive to obtain sufficient data for each new environment.
We propose a meta reinforcement learning (MRL) method to improve the agent's generalization capabilities.
arXiv Detail & Related papers (2020-08-28T02:57:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.