Meta Reinforcement Learning-Based Lane Change Strategy for Autonomous
Vehicles
- URL: http://arxiv.org/abs/2008.12451v1
- Date: Fri, 28 Aug 2020 02:57:11 GMT
- Title: Meta Reinforcement Learning-Based Lane Change Strategy for Autonomous
Vehicles
- Authors: Fei Ye, Pin Wang, Ching-Yao Chan and Jiucai Zhang
- Abstract summary: Supervised learning algorithms can generalize to new environments by training on a large amount of labeled data.
It can be often impractical or cost-prohibitive to obtain sufficient data for each new environment.
We propose a meta reinforcement learning (MRL) method to improve the agent's generalization capabilities.
- Score: 11.180588185127892
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recent advances in supervised learning and reinforcement learning have
provided new opportunities to apply related methodologies to automated driving.
However, there are still challenges to achieve automated driving maneuvers in
dynamically changing environments. Supervised learning algorithms such as
imitation learning can generalize to new environments by training on a large
amount of labeled data, however, it can be often impractical or
cost-prohibitive to obtain sufficient data for each new environment. Although
reinforcement learning methods can mitigate this data-dependency issue by
training the agent in a trial-and-error way, they still need to re-train
policies from scratch when adapting to new environments. In this paper, we thus
propose a meta reinforcement learning (MRL) method to improve the agent's
generalization capabilities to make automated lane-changing maneuvers at
different traffic environments, which are formulated as different traffic
congestion levels. Specifically, we train the model at light to moderate
traffic densities and test it at a new heavy traffic density condition. We use
both collision rate and success rate to quantify the safety and effectiveness
of the proposed model. A benchmark model is developed based on a pretraining
method, which uses the same network structure and training tasks as our
proposed model for fair comparison. The simulation results shows that the
proposed method achieves an overall success rate up to 20% higher than the
benchmark model when it is generalized to the new environment of heavy traffic
density. The collision rate is also reduced by up to 18% than the benchmark
model. Finally, the proposed model shows more stable and efficient
generalization capabilities adapting to the new environment, and it can achieve
100% successful rate and 0% collision rate with only a few steps of gradient
updates.
Related papers
- Mitigating Covariate Shift in Imitation Learning for Autonomous Vehicles Using Latent Space Generative World Models [60.87795376541144]
A world model is a neural network capable of predicting an agent's next state given past states and actions.
During end-to-end training, our policy learns how to recover from errors by aligning with states observed in human demonstrations.
We present qualitative and quantitative results, demonstrating significant improvements upon prior state of the art in closed-loop testing.
arXiv Detail & Related papers (2024-09-25T06:48:25Z) - Traffic expertise meets residual RL: Knowledge-informed model-based residual reinforcement learning for CAV trajectory control [1.5361702135159845]
This paper introduces a knowledge-informed model-based residual reinforcement learning framework.
It integrates traffic expert knowledge into a virtual environment model, employing the Intelligent Driver Model (IDM) for basic dynamics and neural networks for residual dynamics.
We propose a novel strategy that combines traditional control methods with residual RL, facilitating efficient learning and policy optimization without the need to learn from scratch.
arXiv Detail & Related papers (2024-08-30T16:16:57Z) - Continual Learning for Adaptable Car-Following in Dynamic Traffic Environments [16.587883982785]
The continual evolution of autonomous driving technology requires car-following models that can adapt to diverse and dynamic traffic environments.
Traditional learning-based models often suffer from performance degradation when encountering unseen traffic patterns due to a lack of continual learning capabilities.
This paper proposes a novel car-following model based on continual learning that addresses this limitation.
arXiv Detail & Related papers (2024-07-17T06:32:52Z) - MetaFollower: Adaptable Personalized Autonomous Car Following [63.90050686330677]
We propose an adaptable personalized car-following framework - MetaFollower.
We first utilize Model-Agnostic Meta-Learning (MAML) to extract common driving knowledge from various CF events.
We additionally combine Long Short-Term Memory (LSTM) and Intelligent Driver Model (IDM) to reflect temporal heterogeneity with high interpretability.
arXiv Detail & Related papers (2024-06-23T15:30:40Z) - Data-Efficient Task Generalization via Probabilistic Model-based Meta
Reinforcement Learning [58.575939354953526]
PACOH-RL is a novel model-based Meta-Reinforcement Learning (Meta-RL) algorithm designed to efficiently adapt control policies to changing dynamics.
Existing Meta-RL methods require abundant meta-learning data, limiting their applicability in settings such as robotics.
Our experiment results demonstrate that PACOH-RL outperforms model-based RL and model-based Meta-RL baselines in adapting to new dynamic conditions.
arXiv Detail & Related papers (2023-11-13T18:51:57Z) - Unsupervised Domain Adaptation for Self-Driving from Past Traversal
Features [69.47588461101925]
We propose a method to adapt 3D object detectors to new driving environments.
Our approach enhances LiDAR-based detection models using spatial quantized historical features.
Experiments on real-world datasets demonstrate significant improvements.
arXiv Detail & Related papers (2023-09-21T15:00:31Z) - ADLight: A Universal Approach of Traffic Signal Control with Augmented
Data Using Reinforcement Learning [3.3458830284045065]
We propose a novel reinforcement learning approach with augmented data (ADLight) to train a universal model for intersections with different structures.
A new data augmentation method named textitmovement shuffle is developed to improve the generalization performance.
The results show that the performance of our approach is close to the models trained in a single environment directly.
arXiv Detail & Related papers (2022-10-24T16:21:48Z) - Efficient Model-based Multi-agent Reinforcement Learning via Optimistic
Equilibrium Computation [93.52573037053449]
H-MARL (Hallucinated Multi-Agent Reinforcement Learning) learns successful equilibrium policies after a few interactions with the environment.
We demonstrate our approach experimentally on an autonomous driving simulation benchmark.
arXiv Detail & Related papers (2022-03-14T17:24:03Z) - AdaPool: A Diurnal-Adaptive Fleet Management Framework using Model-Free
Deep Reinforcement Learning and Change Point Detection [34.77250498401055]
This paper introduces an adaptive model-free deep reinforcement approach that can recognize and adapt to the diurnal patterns in the ride-sharing environment with car-pooling.
In addition to the adaptation logic in dispatching, this paper also proposes a dynamic, demand-aware vehicle-passenger matching and route planning framework.
arXiv Detail & Related papers (2021-04-01T02:14:01Z) - MetaVIM: Meta Variationally Intrinsic Motivated Reinforcement Learning for Decentralized Traffic Signal Control [54.162449208797334]
Traffic signal control aims to coordinate traffic signals across intersections to improve the traffic efficiency of a district or a city.
Deep reinforcement learning (RL) has been applied to traffic signal control recently and demonstrated promising performance where each traffic signal is regarded as an agent.
We propose a novel Meta Variationally Intrinsic Motivated (MetaVIM) RL method to learn the decentralized policy for each intersection that considers neighbor information in a latent way.
arXiv Detail & Related papers (2021-01-04T03:06:08Z) - Bayesian Meta-reinforcement Learning for Traffic Signal Control [5.025654873456756]
We propose a novel value-based Bayesian meta-reinforcement learning framework BM-DQN to robustly speed up the learning process in new scenarios.
The experiments on restricted 2D navigation and traffic signal control show that our proposed framework adapts more quickly and robustly in new scenarios than previous methods.
arXiv Detail & Related papers (2020-10-01T01:15:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.