Integration of Reinforcement Learning Based Behavior Planning With
Sampling Based Motion Planning for Automated Driving
- URL: http://arxiv.org/abs/2304.08280v1
- Date: Mon, 17 Apr 2023 13:49:55 GMT
- Title: Integration of Reinforcement Learning Based Behavior Planning With
Sampling Based Motion Planning for Automated Driving
- Authors: Marvin Klimke, Benjamin V\"olz, Michael Buchholz
- Abstract summary: We propose a method to employ a trained deep reinforcement learning policy for dedicated high-level behavior planning.
To the best of our knowledge, this work is the first to apply deep reinforcement learning in this manner.
- Score: 0.5801044612920815
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Reinforcement learning has received high research interest for developing
planning approaches in automated driving. Most prior works consider the
end-to-end planning task that yields direct control commands and rarely deploy
their algorithm to real vehicles. In this work, we propose a method to employ a
trained deep reinforcement learning policy for dedicated high-level behavior
planning. By populating an abstract objective interface, established motion
planning algorithms can be leveraged, which derive smooth and drivable
trajectories. Given the current environment model, we propose to use a built-in
simulator to predict the traffic scene for a given horizon into the future. The
behavior of automated vehicles in mixed traffic is determined by querying the
learned policy. To the best of our knowledge, this work is the first to apply
deep reinforcement learning in this manner, and as such lacks a
state-of-the-art benchmark. Thus, we validate the proposed approach by
comparing an idealistic single-shot plan with cyclic replanning through the
learned policy. Experiments with a real testing vehicle on proving grounds
demonstrate the potential of our approach to shrink the simulation to real
world gap of deep reinforcement learning based planning approaches. Additional
simulative analyses reveal that more complex multi-agent maneuvers can be
managed by employing the cycling replanning approach.
Related papers
- HOPE: A Reinforcement Learning-based Hybrid Policy Path Planner for Diverse Parking Scenarios [24.25807334214834]
We introduce Hybrid pOlicy Path plannEr (HOPE) to handle diverse and complex parking scenarios.
HOPE integrates a reinforcement learning agent with Reeds-Shepp curves, enabling effective planning across diverse scenarios.
We propose a criterion for categorizing the difficulty level of parking scenarios based on space and obstacle distribution.
arXiv Detail & Related papers (2024-05-31T02:17:51Z) - Speeding Up Path Planning via Reinforcement Learning in MCTS for Automated Parking [3.750010944080163]
We propose a reinforcement learning pipeline with a Monte Carlo tree search under the path planning framework.
By iteratively learning the value of a state, we are able to model a value estimator and a policy generator for given states.
arXiv Detail & Related papers (2024-03-25T22:21:23Z) - LLM-Assist: Enhancing Closed-Loop Planning with Language-Based Reasoning [65.86754998249224]
We develop a novel hybrid planner that leverages a conventional rule-based planner in conjunction with an LLM-based planner.
Our approach navigates complex scenarios which existing planners struggle with, produces well-reasoned outputs while also remaining grounded through working alongside the rule-based approach.
arXiv Detail & Related papers (2023-12-30T02:53:45Z) - Generating Useful Accident-Prone Driving Scenarios via a Learned Traffic
Prior [135.78858513845233]
STRIVE is a method to automatically generate challenging scenarios that cause a given planner to produce undesirable behavior, like collisions.
To maintain scenario plausibility, the key idea is to leverage a learned model of traffic motion in the form of a graph-based conditional VAE.
A subsequent optimization is used to find a "solution" to the scenario, ensuring it is useful to improve the given planner.
arXiv Detail & Related papers (2021-12-09T18:03:27Z) - Learning Interactive Driving Policies via Data-driven Simulation [125.97811179463542]
Data-driven simulators promise high data-efficiency for driving policy learning.
Small underlying datasets often lack interesting and challenging edge cases for learning interactive driving.
We propose a simulation method that uses in-painted ado vehicles for learning robust driving policies.
arXiv Detail & Related papers (2021-11-23T20:14:02Z) - Learning to drive from a world on rails [78.28647825246472]
We learn an interactive vision-based driving policy from pre-recorded driving logs via a model-based approach.
A forward model of the world supervises a driving policy that predicts the outcome of any potential driving trajectory.
Our method ranks first on the CARLA leaderboard, attaining a 25% higher driving score while using 40 times less data.
arXiv Detail & Related papers (2021-05-03T05:55:30Z) - Real-world Ride-hailing Vehicle Repositioning using Deep Reinforcement
Learning [52.2663102239029]
We present a new practical framework based on deep reinforcement learning and decision-time planning for real-world vehicle on idle-hailing platforms.
Our approach learns ride-based state-value function using a batch training algorithm with deep value.
We benchmark our algorithm with baselines in a ride-hailing simulation environment to demonstrate its superiority in improving income efficiency.
arXiv Detail & Related papers (2021-03-08T05:34:05Z) - Experience-Based Heuristic Search: Robust Motion Planning with Deep
Q-Learning [0.0]
We show how experiences in the form of a Deep Q-Network can be integrated as optimal policy in a search algorithm.
Our method may encourage further investigation of the applicability of reinforcement-learning-based planning in the field of self-driving vehicles.
arXiv Detail & Related papers (2021-02-05T12:08:11Z) - Affordance-based Reinforcement Learning for Urban Driving [3.507764811554557]
We propose a deep reinforcement learning framework to learn optimal control policy using waypoints and low-dimensional visual representations.
We demonstrate that our agents when trained from scratch learn the tasks of lane-following, driving around inter-sections as well as stopping in front of other actors or traffic lights even in the dense traffic setting.
arXiv Detail & Related papers (2021-01-15T05:21:25Z) - Planning on the fast lane: Learning to interact using attention
mechanisms in path integral inverse reinforcement learning [20.435909887810165]
General-purpose trajectory planning algorithms for automated driving utilize complex reward functions.
Deep learning approaches have been successfully applied to predict local situation-dependent reward functions.
We present a neural network architecture that uses a policy attention mechanism to generate a low-dimensional context vector.
arXiv Detail & Related papers (2020-07-11T15:25:44Z) - The Importance of Prior Knowledge in Precise Multimodal Prediction [71.74884391209955]
Roads have well defined geometries, topologies, and traffic rules.
In this paper we propose to incorporate structured priors as a loss function.
We demonstrate the effectiveness of our approach on real-world self-driving datasets.
arXiv Detail & Related papers (2020-06-04T03:56:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.