Driver Dojo: A Benchmark for Generalizable Reinforcement Learning for
Autonomous Driving
- URL: http://arxiv.org/abs/2207.11432v1
- Date: Sat, 23 Jul 2022 06:29:43 GMT
- Title: Driver Dojo: A Benchmark for Generalizable Reinforcement Learning for
Autonomous Driving
- Authors: Sebastian Rietsch, Shih-Yuan Huang, Georgios Kontes, Axel Plinge,
Christopher Mutschler
- Abstract summary: We propose a benchmark for generalizable reinforcement learning for autonomous driving.
Our application-oriented benchmark enables a better understanding of the impact of design decisions.
Our benchmark aims to encourage researchers to propose solutions that are able to successfully generalize across scenarios.
- Score: 1.496194593196997
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Reinforcement learning (RL) has shown to reach super human-level performance
across a wide range of tasks. However, unlike supervised machine learning,
learning strategies that generalize well to a wide range of situations remains
one of the most challenging problems for real-world RL. Autonomous driving (AD)
provides a multi-faceted experimental field, as it is necessary to learn the
correct behavior over many variations of road layouts and large distributions
of possible traffic situations, including individual driver personalities and
hard-to-predict traffic events. In this paper we propose a challenging
benchmark for generalizable RL for AD based on a configurable, flexible, and
performant code base. Our benchmark uses a catalog of randomized scenario
generators, including multiple mechanisms for road layout and traffic
variations, different numerical and visual observation types, distinct action
spaces, diverse vehicle models, and allows for use under static scenario
definitions. In addition to purely algorithmic insights, our
application-oriented benchmark also enables a better understanding of the
impact of design decisions such as action and observation space on the
generalizability of policies. Our benchmark aims to encourage researchers to
propose solutions that are able to successfully generalize across scenarios, a
task in which current RL methods fail. The code for the benchmark is available
at https://github.com/seawee1/driver-dojo.
Related papers
- ODRL: A Benchmark for Off-Dynamics Reinforcement Learning [59.72217833812439]
We introduce ODRL, the first benchmark tailored for evaluating off-dynamics RL methods.
ODRL contains four experimental settings where the source and target domains can be either online or offline.
We conduct extensive benchmarking experiments, which show that no method has universal advantages across varied dynamics shifts.
arXiv Detail & Related papers (2024-10-28T05:29:38Z) - Generalizing Cooperative Eco-driving via Multi-residual Task Learning [6.864745785996583]
Multi-residual Task Learning (MRTL) is a generic learning framework based on multi-task learning.
MRTL decomposes control into nominal components that are effectively solved by conventional control methods and residual terms.
We employ MRTL for fleet-level emission reduction in mixed traffic using autonomous vehicles as a means of system control.
arXiv Detail & Related papers (2024-03-07T05:25:34Z) - Action and Trajectory Planning for Urban Autonomous Driving with
Hierarchical Reinforcement Learning [1.3397650653650457]
We propose an action and trajectory planner using Hierarchical Reinforcement Learning (atHRL) method.
We empirically verify the efficacy of atHRL through extensive experiments in complex urban driving scenarios.
arXiv Detail & Related papers (2023-06-28T07:11:02Z) - Comprehensive Training and Evaluation on Deep Reinforcement Learning for
Automated Driving in Various Simulated Driving Maneuvers [0.4241054493737716]
This study implements, evaluating, and comparing the two DRL algorithms, Deep Q-networks (DQN) and Trust Region Policy Optimization (TRPO)
Models trained on the designed ComplexRoads environment can adapt well to other driving maneuvers with promising overall performance.
arXiv Detail & Related papers (2023-06-20T11:41:01Z) - FastRLAP: A System for Learning High-Speed Driving via Deep RL and
Autonomous Practicing [71.76084256567599]
We present a system that enables an autonomous small-scale RC car to drive aggressively from visual observations using reinforcement learning (RL)
Our system, FastRLAP (faster lap), trains autonomously in the real world, without human interventions, and without requiring any simulation or expert demonstrations.
The resulting policies exhibit emergent aggressive driving skills, such as timing braking and acceleration around turns and avoiding areas which impede the robot's motion, approaching the performance of a human driver using a similar first-person interface over the course of training.
arXiv Detail & Related papers (2023-04-19T17:33:47Z) - Visual Exemplar Driven Task-Prompting for Unified Perception in
Autonomous Driving [100.3848723827869]
We present an effective multi-task framework, VE-Prompt, which introduces visual exemplars via task-specific prompting.
Specifically, we generate visual exemplars based on bounding boxes and color-based markers, which provide accurate visual appearances of target categories.
We bridge transformer-based encoders and convolutional layers for efficient and accurate unified perception in autonomous driving.
arXiv Detail & Related papers (2023-03-03T08:54:06Z) - Automated Reinforcement Learning (AutoRL): A Survey and Open Problems [92.73407630874841]
Automated Reinforcement Learning (AutoRL) involves not only standard applications of AutoML but also includes additional challenges unique to RL.
We provide a common taxonomy, discuss each area in detail and pose open problems which would be of interest to researchers going forward.
arXiv Detail & Related papers (2022-01-11T12:41:43Z) - MetaDrive: Composing Diverse Driving Scenarios for Generalizable
Reinforcement Learning [25.191567110519866]
We develop a new driving simulation platform called MetaDrive for the study of reinforcement learning algorithms.
Based on MetaDrive, we construct a variety of RL tasks and baselines in both single-agent and multi-agent settings.
arXiv Detail & Related papers (2021-09-26T18:34:55Z) - End-to-End Intersection Handling using Multi-Agent Deep Reinforcement
Learning [63.56464608571663]
Navigating through intersections is one of the main challenging tasks for an autonomous vehicle.
In this work, we focus on the implementation of a system able to navigate through intersections where only traffic signs are provided.
We propose a multi-agent system using a continuous, model-free Deep Reinforcement Learning algorithm used to train a neural network for predicting both the acceleration and the steering angle at each time step.
arXiv Detail & Related papers (2021-04-28T07:54:40Z) - Investigating Value of Curriculum Reinforcement Learning in Autonomous
Driving Under Diverse Road and Weather Conditions [0.0]
This paper focuses on investigating the value of curriculum reinforcement learning in autonomous driving applications.
We setup several different driving scenarios in a realistic driving simulator, with varying road complexity and weather conditions.
Results show that curriculum RL can yield significant gains in complex driving tasks, both in terms of driving performance and sample complexity.
arXiv Detail & Related papers (2021-03-14T12:05:05Z) - Trajectory-wise Multiple Choice Learning for Dynamics Generalization in
Reinforcement Learning [137.39196753245105]
We present a new model-based reinforcement learning algorithm that learns a multi-headed dynamics model for dynamics generalization.
We incorporate context learning, which encodes dynamics-specific information from past experiences into the context latent vector.
Our method exhibits superior zero-shot generalization performance across a variety of control tasks, compared to state-of-the-art RL methods.
arXiv Detail & Related papers (2020-10-26T03:20:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.