Carl-Lead: Lidar-based End-to-End Autonomous Driving with Contrastive
Deep Reinforcement Learning
- URL: http://arxiv.org/abs/2109.08473v1
- Date: Fri, 17 Sep 2021 11:24:10 GMT
- Title: Carl-Lead: Lidar-based End-to-End Autonomous Driving with Contrastive
Deep Reinforcement Learning
- Authors: Peide Cai, Sukai Wang, Hengli Wang, Ming Liu
- Abstract summary: We use deep reinforcement learning (DRL) to train lidar-based end-to-end driving policies.
In this work, we use DRL to train lidar-based end-to-end driving policies that naturally consider imperfect partial observations.
Our method achieves higher success rates than the state-of-the-art (SOTA) lidar-based end-to-end driving network.
- Score: 10.040113551761792
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Autonomous driving in urban crowds at unregulated intersections is
challenging, where dynamic occlusions and uncertain behaviors of other vehicles
should be carefully considered. Traditional methods are heuristic and based on
hand-engineered rules and parameters, but scale poorly in new situations.
Therefore, they require high labor cost to design and maintain rules in all
foreseeable scenarios. Recently, deep reinforcement learning (DRL) has shown
promising results in urban driving scenarios. However, DRL is known to be
sample inefficient, and most previous works assume perfect observations such as
ground-truth locations and motions of vehicles without considering noises and
occlusions, which might be a too strong assumption for policy deployment. In
this work, we use DRL to train lidar-based end-to-end driving policies that
naturally consider imperfect partial observations. We further use unsupervised
contrastive representation learning as an auxiliary task to improve the sample
efficiency. The comparative evaluation results reveal that our method achieves
higher success rates than the state-of-the-art (SOTA) lidar-based end-to-end
driving network, better trades off safety and efficiency than the carefully
tuned rule-based method, and generalizes better to new scenarios than the
baselines. Demo videos are available at https://caipeide.github.io/carl-lead/.
Related papers
- Learning Realistic Traffic Agents in Closed-loop [36.38063449192355]
Reinforcement learning (RL) can train traffic agents to avoid infractions, but using RL alone results in unhuman-like driving behaviors.
We propose Reinforcing Traffic Rules (RTR) to match expert demonstrations under a traffic compliance constraint.
Our experiments show that RTR learns more realistic and generalizable traffic simulation policies.
arXiv Detail & Related papers (2023-11-02T16:55:23Z) - Action and Trajectory Planning for Urban Autonomous Driving with
Hierarchical Reinforcement Learning [1.3397650653650457]
We propose an action and trajectory planner using Hierarchical Reinforcement Learning (atHRL) method.
We empirically verify the efficacy of atHRL through extensive experiments in complex urban driving scenarios.
arXiv Detail & Related papers (2023-06-28T07:11:02Z) - Rethinking Closed-loop Training for Autonomous Driving [82.61418945804544]
We present the first empirical study which analyzes the effects of different training benchmark designs on the success of learning agents.
We propose trajectory value learning (TRAVL), an RL-based driving agent that performs planning with multistep look-ahead.
Our experiments show that TRAVL can learn much faster and produce safer maneuvers compared to all the baselines.
arXiv Detail & Related papers (2023-06-27T17:58:39Z) - DenseLight: Efficient Control for Large-scale Traffic Signals with Dense
Feedback [109.84667902348498]
Traffic Signal Control (TSC) aims to reduce the average travel time of vehicles in a road network.
Most prior TSC methods leverage deep reinforcement learning to search for a control policy.
We propose DenseLight, a novel RL-based TSC method that employs an unbiased reward function to provide dense feedback on policy effectiveness.
arXiv Detail & Related papers (2023-06-13T05:58:57Z) - DQ-GAT: Towards Safe and Efficient Autonomous Driving with Deep
Q-Learning and Graph Attention Networks [12.714551756377265]
Traditional planning methods are largely rule-based and scale poorly in complex dynamic scenarios.
We propose DQ-GAT to achieve scalable and proactive autonomous driving.
Our method can better trade-off safety and efficiency in both seen and unseen scenarios.
arXiv Detail & Related papers (2021-08-11T04:55:23Z) - Vision-Based Autonomous Car Racing Using Deep Imitative Reinforcement
Learning [13.699336307578488]
Deep imitative reinforcement learning approach (DIRL) achieves agile autonomous racing using visual inputs.
We validate our algorithm both in a high-fidelity driving simulation and on a real-world 1/20-scale RC-car with limited onboard computation.
arXiv Detail & Related papers (2021-07-18T00:00:48Z) - Learning to drive from a world on rails [78.28647825246472]
We learn an interactive vision-based driving policy from pre-recorded driving logs via a model-based approach.
A forward model of the world supervises a driving policy that predicts the outcome of any potential driving trajectory.
Our method ranks first on the CARLA leaderboard, attaining a 25% higher driving score while using 40 times less data.
arXiv Detail & Related papers (2021-05-03T05:55:30Z) - End-to-End Intersection Handling using Multi-Agent Deep Reinforcement
Learning [63.56464608571663]
Navigating through intersections is one of the main challenging tasks for an autonomous vehicle.
In this work, we focus on the implementation of a system able to navigate through intersections where only traffic signs are provided.
We propose a multi-agent system using a continuous, model-free Deep Reinforcement Learning algorithm used to train a neural network for predicting both the acceleration and the steering angle at each time step.
arXiv Detail & Related papers (2021-04-28T07:54:40Z) - Injecting Knowledge in Data-driven Vehicle Trajectory Predictors [82.91398970736391]
Vehicle trajectory prediction tasks have been commonly tackled from two perspectives: knowledge-driven or data-driven.
In this paper, we propose to learn a "Realistic Residual Block" (RRB) which effectively connects these two perspectives.
Our proposed method outputs realistic predictions by confining the residual range and taking into account its uncertainty.
arXiv Detail & Related papers (2021-03-08T16:03:09Z) - MetaVIM: Meta Variationally Intrinsic Motivated Reinforcement Learning for Decentralized Traffic Signal Control [54.162449208797334]
Traffic signal control aims to coordinate traffic signals across intersections to improve the traffic efficiency of a district or a city.
Deep reinforcement learning (RL) has been applied to traffic signal control recently and demonstrated promising performance where each traffic signal is regarded as an agent.
We propose a novel Meta Variationally Intrinsic Motivated (MetaVIM) RL method to learn the decentralized policy for each intersection that considers neighbor information in a latent way.
arXiv Detail & Related papers (2021-01-04T03:06:08Z) - Behavior Planning at Urban Intersections through Hierarchical
Reinforcement Learning [25.50973559614565]
In this work, we propose a behavior planning structure based on reinforcement learning (RL) which is capable of performing autonomous vehicle behavior planning with a hierarchical structure in simulated urban environments.
Our algorithms can perform better than rule-based methods for elective decisions such as when to turn left between vehicles approaching from the opposite direction or possible lane-change when approaching an intersection due to lane blockage or delay in front of the ego car.
Results also show that the proposed method converges to an optimal policy faster than traditional RL methods.
arXiv Detail & Related papers (2020-11-09T19:23:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.