Learning hierarchical behavior and motion planning for autonomous
driving
- URL: http://arxiv.org/abs/2005.03863v1
- Date: Fri, 8 May 2020 05:34:55 GMT
- Title: Learning hierarchical behavior and motion planning for autonomous
driving
- Authors: Jingke Wang, Yue Wang, Dongkun Zhang, Yezhou Yang, Rong Xiong
- Abstract summary: We introduce hierarchical behavior and motion planning (HBMP) to explicitly model the behavior in learning-based solution.
We transform HBMP problem by integrating a classical sampling-based motion planner.
In addition, we propose a sharable representation for input sensory data across simulation platforms and real-world environment.
- Score: 32.78069835190924
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Learning-based driving solution, a new branch for autonomous driving, is
expected to simplify the modeling of driving by learning the underlying
mechanisms from data. To improve the tactical decision-making for
learning-based driving solution, we introduce hierarchical behavior and motion
planning (HBMP) to explicitly model the behavior in learning-based solution.
Due to the coupled action space of behavior and motion, it is challenging to
solve HBMP problem using reinforcement learning (RL) for long-horizon driving
tasks. We transform HBMP problem by integrating a classical sampling-based
motion planner, of which the optimal cost is regarded as the rewards for
high-level behavior learning. As a result, this formulation reduces action
space and diversifies the rewards without losing the optimality of HBMP. In
addition, we propose a sharable representation for input sensory data across
simulation platforms and real-world environment, so that models trained in a
fast event-based simulator, SUMO, can be used to initialize and accelerate the
RL training in a dynamics based simulator, CARLA. Experimental results
demonstrate the effectiveness of the method. Besides, the model is successfully
transferred to the real-world, validating the generalization capability.
Related papers
- Hierarchical and Decoupled BEV Perception Learning Framework for Autonomous Driving [52.808273563372126]
This paper proposes a novel hierarchical BEV perception paradigm, aiming to provide a library of fundamental perception modules and user-friendly graphical interface.
We conduct the Pretrain-Finetune strategy to effectively utilize large scale public datasets and streamline development processes.
We also present a Multi-Module Learning (MML) approach, enhancing performance through synergistic and iterative training of multiple models.
arXiv Detail & Related papers (2024-07-17T11:17:20Z) - Towards Human-Like Driving: Active Inference in Autonomous Vehicle Control [0.5437298646956507]
This paper presents a novel approach to Autonomous Vehicle (AV) control through the application of active inference.
Active inference is a theory derived from neuroscience that conceptualizes the brain as a predictive machine.
Our method integrates active inference with deep learning to manage lateral control in AVs, enabling them to perform lane following maneuvers within a simulated urban environment.
arXiv Detail & Related papers (2024-07-10T14:08:27Z) - Waymax: An Accelerated, Data-Driven Simulator for Large-Scale Autonomous
Driving Research [76.93956925360638]
Waymax is a new data-driven simulator for autonomous driving in multi-agent scenes.
It runs entirely on hardware accelerators such as TPUs/GPUs and supports in-graph simulation for training.
We benchmark a suite of popular imitation and reinforcement learning algorithms with ablation studies on different design decisions.
arXiv Detail & Related papers (2023-10-12T20:49:15Z) - Using Implicit Behavior Cloning and Dynamic Movement Primitive to Facilitate Reinforcement Learning for Robot Motion Planning [3.16488279864227]
Reinforcement learning (RL) for motion planning of robots suffers from low efficiency in terms of slow training speed and poor generalizability.
We propose a novel RL-based framework that uses implicit behavior cloning (IBC) and dynamic movement primitive (DMP) to improve the training speed and generalizability of an off-policy RL agent.
arXiv Detail & Related papers (2023-07-29T19:46:09Z) - FastRLAP: A System for Learning High-Speed Driving via Deep RL and
Autonomous Practicing [71.76084256567599]
We present a system that enables an autonomous small-scale RC car to drive aggressively from visual observations using reinforcement learning (RL)
Our system, FastRLAP (faster lap), trains autonomously in the real world, without human interventions, and without requiring any simulation or expert demonstrations.
The resulting policies exhibit emergent aggressive driving skills, such as timing braking and acceleration around turns and avoiding areas which impede the robot's motion, approaching the performance of a human driver using a similar first-person interface over the course of training.
arXiv Detail & Related papers (2023-04-19T17:33:47Z) - Tackling Real-World Autonomous Driving using Deep Reinforcement Learning [63.3756530844707]
In this work, we propose a model-free Deep Reinforcement Learning Planner training a neural network that predicts acceleration and steering angle.
In order to deploy the system on board the real self-driving car, we also develop a module represented by a tiny neural network.
arXiv Detail & Related papers (2022-07-05T16:33:20Z) - Accelerated Policy Learning with Parallel Differentiable Simulation [59.665651562534755]
We present a differentiable simulator and a new policy learning algorithm (SHAC)
Our algorithm alleviates problems with local minima through a smooth critic function.
We show substantial improvements in sample efficiency and wall-clock time over state-of-the-art RL and differentiable simulation-based algorithms.
arXiv Detail & Related papers (2022-04-14T17:46:26Z) - Spatiotemporal Costmap Inference for MPC via Deep Inverse Reinforcement
Learning [27.243603228431564]
We propose a new IRLRL algorithm that learns a goal-conditionedtemporal reward function.
The resulting costmap is used by Model Predictive Controllers (MPCs) to perform a task.
arXiv Detail & Related papers (2022-01-17T17:36:29Z) - A Probabilistic Framework for Imitating Human Race Driver Behavior [31.524303667746643]
We propose Probabilistic Modeling of Driver behavior (ProMoD), a modular framework which splits the task of driver behavior modeling into multiple modules.
A global target trajectory distribution is learned with Probabilistic Movement Primitives, clothoids are utilized for local path generation, and the corresponding choice of actions is performed by a neural network.
Experiments in a simulated car racing setting show considerable advantages in imitation accuracy and robustness compared to other imitation learning algorithms.
arXiv Detail & Related papers (2020-01-22T20:06:38Z) - Information Theoretic Model Predictive Q-Learning [64.74041985237105]
We present a novel theoretical connection between information theoretic MPC and entropy regularized RL.
We develop a Q-learning algorithm that can leverage biased models.
arXiv Detail & Related papers (2019-12-31T00:29:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.