Towards Robust On-Ramp Merging via Augmented Multimodal Reinforcement
Learning
- URL: http://arxiv.org/abs/2208.07307v1
- Date: Thu, 21 Jul 2022 16:34:57 GMT
- Title: Towards Robust On-Ramp Merging via Augmented Multimodal Reinforcement
Learning
- Authors: Gaurav Bagwe, Jian Li, Xiaoyong Yuan, Lan Zhang
- Abstract summary: We present a novel approach for Robust on-ramp merge of CAVs via Augmented and Multi-modal Reinforcement Learning.
Specifically, we formulate the on-ramp merging problem as a Markov decision process (MDP) by taking driving safety, comfort driving behavior, and traffic efficiency into account.
To provide reliable merging maneuvers, we simultaneously leverage BSM and surveillance images for multi-modal observation.
- Score: 9.48157144651867
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Despite the success of AI-enabled onboard perception, on-ramp merging has
been one of the main challenges for autonomous driving. Due to limited sensing
range of onboard sensors, a merging vehicle can hardly observe main road
conditions and merge properly. By leveraging the wireless communications
between connected and automated vehicles (CAVs), a merging CAV has potential to
proactively obtain the intentions of nearby vehicles. However, CAVs can be
prone to inaccurate observations, such as the noisy basic safety messages (BSM)
and poor quality surveillance images. In this paper, we present a novel
approach for Robust on-ramp merge of CAVs via Augmented and Multi-modal
Reinforcement Learning, named by RAMRL. Specifically, we formulate the on-ramp
merging problem as a Markov decision process (MDP) by taking driving safety,
comfort driving behavior, and traffic efficiency into account. To provide
reliable merging maneuvers, we simultaneously leverage BSM and surveillance
images for multi-modal observation, which is used to learn a policy model
through proximal policy optimization (PPO). Moreover, to improve data
efficiency and provide better generalization performance, we train the policy
model with augmented data (e.g., noisy BSM and noisy surveillance images).
Extensive experiments are conducted with Simulation of Urban MObility (SUMO)
platform under two typical merging scenarios. Experimental results demonstrate
the effectiveness and efficiency of our robust on-ramp merging design.
Related papers
- Towards Interactive and Learnable Cooperative Driving Automation: a Large Language Model-Driven Decision-Making Framework [79.088116316919]
Connected Autonomous Vehicles (CAVs) have begun to open road testing around the world, but their safety and efficiency performance in complex scenarios is still not satisfactory.
This paper proposes CoDrivingLLM, an interactive and learnable LLM-driven cooperative driving framework.
arXiv Detail & Related papers (2024-09-19T14:36:00Z) - Reinforcement Learning with Latent State Inference for Autonomous On-ramp Merging under Observation Delay [6.0111084468944]
We introduce the Lane-keeping, Lane-changing with Latent-state Inference and Safety Controller (L3IS) agent.
L3IS is designed to perform the on-ramp merging task safely without comprehensive knowledge about surrounding vehicles' intents or driving styles.
We present an augmentation of this agent called AL3IS that accounts for observation delays, allowing the agent to make more robust decisions in real-world environments.
arXiv Detail & Related papers (2024-03-18T15:02:46Z) - Empowering Autonomous Driving with Large Language Models: A Safety Perspective [82.90376711290808]
This paper explores the integration of Large Language Models (LLMs) into Autonomous Driving systems.
LLMs are intelligent decision-makers in behavioral planning, augmented with a safety verifier shield for contextual safety learning.
We present two key studies in a simulated environment: an adaptive LLM-conditioned Model Predictive Control (MPC) and an LLM-enabled interactive behavior planning scheme with a state machine.
arXiv Detail & Related papers (2023-11-28T03:13:09Z) - Model-Based Reinforcement Learning with Isolated Imaginations [61.67183143982074]
We propose Iso-Dream++, a model-based reinforcement learning approach.
We perform policy optimization based on the decoupled latent imaginations.
This enables long-horizon visuomotor control tasks to benefit from isolating mixed dynamics sources in the wild.
arXiv Detail & Related papers (2023-03-27T02:55:56Z) - Shared Information-Based Safe And Efficient Behavior Planning For
Connected Autonomous Vehicles [6.896682830421197]
We design an integrated information sharing and safe multi-agent reinforcement learning framework for connected autonomous vehicles.
We first use weight pruned convolutional neural networks (CNN) to process the raw image and point cloud LIDAR data locally at each autonomous vehicle.
We then design a safe actor-critic algorithm that utilizes both a vehicle's local observation and the information received via V2V communication.
arXiv Detail & Related papers (2023-02-08T20:31:41Z) - Learning energy-efficient driving behaviors by imitating experts [75.12960180185105]
This paper examines the role of imitation learning in bridging the gap between control strategies and realistic limitations in communication and sensing.
We show that imitation learning can succeed in deriving policies that, if adopted by 5% of vehicles, may boost the energy-efficiency of networks with varying traffic conditions by 15% using only local observations.
arXiv Detail & Related papers (2022-06-28T17:08:31Z) - COOPERNAUT: End-to-End Driving with Cooperative Perception for Networked
Vehicles [54.61668577827041]
We introduce COOPERNAUT, an end-to-end learning model that uses cross-vehicle perception for vision-based cooperative driving.
Our experiments on AutoCastSim suggest that our cooperative perception driving models lead to a 40% improvement in average success rate.
arXiv Detail & Related papers (2022-05-04T17:55:12Z) - Multi-agent Reinforcement Learning for Cooperative Lane Changing of
Connected and Autonomous Vehicles in Mixed Traffic [16.858651125916133]
Lane-changing is a great challenge for autonomous vehicles (AVs) in mixed and dynamic traffic scenarios.
In this paper, we formulate the lane-changing decision making of multiple AVs in a mixed-traffic highway environment as a multi-agent reinforcement learning (MARL) problem.
Our proposed MARL framework consistently outperforms several state-of-the-art benchmarks in terms of efficiency, safety and driver comfort.
arXiv Detail & Related papers (2021-11-11T17:17:24Z) - Learning Interaction-aware Guidance Policies for Motion Planning in
Dense Traffic Scenarios [8.484564880157148]
This paper presents a novel framework for interaction-aware motion planning in dense traffic scenarios.
We propose to learn, via deep Reinforcement Learning (RL), an interaction-aware policy providing global guidance about the cooperativeness of other vehicles.
The learned policy can reason and guide the local optimization-based planner with interactive behavior to pro-actively merge in dense traffic while remaining safe in case the other vehicles do not yield.
arXiv Detail & Related papers (2021-07-09T16:43:12Z) - Transferable Deep Reinforcement Learning Framework for Autonomous
Vehicles with Joint Radar-Data Communications [69.24726496448713]
We propose an intelligent optimization framework based on the Markov Decision Process (MDP) to help the AV make optimal decisions.
We then develop an effective learning algorithm leveraging recent advances of deep reinforcement learning techniques to find the optimal policy for the AV.
We show that the proposed transferable deep reinforcement learning framework reduces the obstacle miss detection probability by the AV up to 67% compared to other conventional deep reinforcement learning approaches.
arXiv Detail & Related papers (2021-05-28T08:45:37Z) - A Multi-Agent Reinforcement Learning Approach For Safe and Efficient
Behavior Planning Of Connected Autonomous Vehicles [21.132777568170702]
We design an information-sharing-based reinforcement learning framework for connected autonomous vehicles.
We show that our approach can improve the CAV system's efficiency in terms of average velocity and comfort.
We construct an obstacle-at-corner scenario to show that the shared vision can help CAVs to observe obstacles earlier and take action to avoid traffic jams.
arXiv Detail & Related papers (2020-03-09T19:15:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.