Integrating Deep Reinforcement Learning with Model-based Path Planners
for Automated Driving
- URL: http://arxiv.org/abs/2002.00434v2
- Date: Tue, 19 May 2020 17:03:49 GMT
- Title: Integrating Deep Reinforcement Learning with Model-based Path Planners
for Automated Driving
- Authors: Ekim Yurtsever, Linda Capito, Keith Redmill, Umit Ozguner
- Abstract summary: We propose a hybrid approach for integrating a path planning pipe into a vision based DRL framework.
In summary, the DRL agent is trained to follow the path planner's waypoints as close as possible.
Experimental results show that the proposed method can plan its path and navigate between randomly chosen origin-destination points.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Automated driving in urban settings is challenging. Human participant
behavior is difficult to model, and conventional, rule-based Automated Driving
Systems (ADSs) tend to fail when they face unmodeled dynamics. On the other
hand, the more recent, end-to-end Deep Reinforcement Learning (DRL) based
model-free ADSs have shown promising results. However, pure learning-based
approaches lack the hard-coded safety measures of model-based controllers. Here
we propose a hybrid approach for integrating a path planning pipe into a vision
based DRL framework to alleviate the shortcomings of both worlds. In summary,
the DRL agent is trained to follow the path planner's waypoints as close as
possible. The agent learns this policy by interacting with the environment. The
reward function contains two major terms: the penalty of straying away from the
path planner and the penalty of having a collision. The latter has precedence
in the form of having a significantly greater numerical value. Experimental
results show that the proposed method can plan its path and navigate between
randomly chosen origin-destination points in CARLA, a dynamic urban simulation
environment. Our code is open-source and available online.
Related papers
- DiFSD: Ego-Centric Fully Sparse Paradigm with Uncertainty Denoising and Iterative Refinement for Efficient End-to-End Autonomous Driving [55.53171248839489]
We propose an ego-centric fully sparse paradigm, named DiFSD, for end-to-end self-driving.
Specifically, DiFSD mainly consists of sparse perception, hierarchical interaction and iterative motion planner.
Experiments conducted on nuScenes dataset demonstrate the superior planning performance and great efficiency of DiFSD.
arXiv Detail & Related papers (2024-09-15T15:55:24Z) - Autonomous Vehicle Controllers From End-to-End Differentiable Simulation [60.05963742334746]
We propose a differentiable simulator and design an analytic policy gradients (APG) approach to training AV controllers.
Our proposed framework brings the differentiable simulator into an end-to-end training loop, where gradients of environment dynamics serve as a useful prior to help the agent learn a more grounded policy.
We find significant improvements in performance and robustness to noise in the dynamics, as well as overall more intuitive human-like handling.
arXiv Detail & Related papers (2024-09-12T11:50:06Z) - Let Hybrid A* Path Planner Obey Traffic Rules: A Deep Reinforcement Learning-Based Planning Framework [0.0]
In this work, we combine low-level algorithms such as the hybrid A* path planning with deep reinforcement learning (DRL) to make high-level decisions.
The hybrid A* planner is able to generate a collision-free trajectory to be executed by a model predictive controller (MPC)
In addition, the DRL algorithm is able to keep the lane change command consistent within a chosen time-period.
arXiv Detail & Related papers (2024-07-01T12:00:10Z) - Planning with Adaptive World Models for Autonomous Driving [50.4439896514353]
Motion planners (MPs) are crucial for safe navigation in complex urban environments.
nuPlan, a recently released MP benchmark, addresses this limitation by augmenting real-world driving logs with closed-loop simulation logic.
We present AdaptiveDriver, a model-predictive control (MPC) based planner that unrolls different world models conditioned on BehaviorNet's predictions.
arXiv Detail & Related papers (2024-06-15T18:53:45Z) - WROOM: An Autonomous Driving Approach for Off-Road Navigation [17.74237088460657]
We design an end-to-end reinforcement learning (RL) system for an autonomous vehicle in off-road environments.
We warm-start the agent by imitating a rule-based controller and utilize Proximal Policy Optimization (PPO) to improve the policy.
We propose a novel simulation environment to replicate off-road driving scenarios and deploy our proposed approach on a real buggy RC car.
arXiv Detail & Related papers (2024-04-12T23:55:59Z) - Action and Trajectory Planning for Urban Autonomous Driving with
Hierarchical Reinforcement Learning [1.3397650653650457]
We propose an action and trajectory planner using Hierarchical Reinforcement Learning (atHRL) method.
We empirically verify the efficacy of atHRL through extensive experiments in complex urban driving scenarios.
arXiv Detail & Related papers (2023-06-28T07:11:02Z) - Causal Imitative Model for Autonomous Driving [85.78593682732836]
We propose Causal Imitative Model (CIM) to address inertia and collision problems.
CIM explicitly discovers the causal model and utilizes it to train the policy.
Our experiments show that our method outperforms previous work in terms of inertia and collision rates.
arXiv Detail & Related papers (2021-12-07T18:59:15Z) - UMBRELLA: Uncertainty-Aware Model-Based Offline Reinforcement Learning
Leveraging Planning [1.1339580074756188]
Offline reinforcement learning (RL) provides a framework for learning decision-making from offline data.
Self-driving vehicles (SDV) learn a policy, which potentially even outperforms the behavior in the sub-optimal data set.
This motivates the use of model-based offline RL approaches, which leverage planning.
arXiv Detail & Related papers (2021-11-22T10:37:52Z) - Learning to drive from a world on rails [78.28647825246472]
We learn an interactive vision-based driving policy from pre-recorded driving logs via a model-based approach.
A forward model of the world supervises a driving policy that predicts the outcome of any potential driving trajectory.
Our method ranks first on the CARLA leaderboard, attaining a 25% higher driving score while using 40 times less data.
arXiv Detail & Related papers (2021-05-03T05:55:30Z) - Modular Deep Reinforcement Learning for Continuous Motion Planning with
Temporal Logic [59.94347858883343]
This paper investigates the motion planning of autonomous dynamical systems modeled by Markov decision processes (MDP)
The novelty is to design an embedded product MDP (EP-MDP) between the LDGBA and the MDP.
The proposed LDGBA-based reward shaping and discounting schemes for the model-free reinforcement learning (RL) only depend on the EP-MDP states.
arXiv Detail & Related papers (2021-02-24T01:11:25Z) - Trajectory Planning for Autonomous Vehicles Using Hierarchical
Reinforcement Learning [21.500697097095408]
Planning safe trajectories under uncertain and dynamic conditions makes the autonomous driving problem significantly complex.
Current sampling-based methods such as Rapidly Exploring Random Trees (RRTs) are not ideal for this problem because of the high computational cost.
We propose a Hierarchical Reinforcement Learning structure combined with a Proportional-Integral-Derivative (PID) controller for trajectory planning.
arXiv Detail & Related papers (2020-11-09T20:49:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.