Safe Model-based Off-policy Reinforcement Learning for Eco-Driving in
Connected and Automated Hybrid Electric Vehicles
- URL: http://arxiv.org/abs/2105.11640v1
- Date: Tue, 25 May 2021 03:41:29 GMT
- Title: Safe Model-based Off-policy Reinforcement Learning for Eco-Driving in
Connected and Automated Hybrid Electric Vehicles
- Authors: Zhaoxuan Zhu, Nicola Pivaro, Shobhit Gupta, Abhishek Gupta and
Marcello Canova
- Abstract summary: This work proposes a Safe Off-policy Model-Based Reinforcement Learning algorithm for the eco-driving problem.
The proposed algorithm leads to a policy with a higher average speed and a better fuel economy compared to the model-free agent.
- Score: 3.5259944260228977
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Connected and Automated Hybrid Electric Vehicles have the potential to reduce
fuel consumption and travel time in real-world driving conditions. The
eco-driving problem seeks to design optimal speed and power usage profiles
based upon look-ahead information from connectivity and advanced mapping
features. Recently, Deep Reinforcement Learning (DRL) has been applied to the
eco-driving problem. While the previous studies synthesize simulators and
model-free DRL to reduce online computation, this work proposes a Safe
Off-policy Model-Based Reinforcement Learning algorithm for the eco-driving
problem. The advantages over the existing literature are three-fold. First, the
combination of off-policy learning and the use of a physics-based model
improves the sample efficiency. Second, the training does not require any
extrinsic rewarding mechanism for constraint satisfaction. Third, the
feasibility of trajectory is guaranteed by using a safe set approximated by
deep generative models.
The performance of the proposed method is benchmarked against a baseline
controller representing human drivers, a previously designed model-free DRL
strategy, and the wait-and-see optimal solution. In simulation, the proposed
algorithm leads to a policy with a higher average speed and a better fuel
economy compared to the model-free agent. Compared to the baseline controller,
the learned strategy reduces the fuel consumption by more than 21\% while
keeping the average speed comparable.
Related papers
- Physics Enhanced Residual Policy Learning (PERPL) for safety cruising in mixed traffic platooning under actuator and communication delay [8.172286651098027]
Linear control models have gained extensive application in vehicle control due to their simplicity, ease of use, and support for stability analysis.
Reinforcement learning (RL) models, on the other hand, offer adaptability but suffer from a lack of interpretability and generalization capabilities.
This paper aims to develop a family of RL-based controllers enhanced by physics-informed policies.
arXiv Detail & Related papers (2024-09-23T23:02:34Z) - EcoFollower: An Environment-Friendly Car Following Model Considering Fuel Consumption [9.42048156323799]
This study introduces EcoFollower, a novel eco-car-following model developed using reinforcement learning (RL) to optimize fuel consumption in car-following scenarios.
The model achieved a significant reduction in fuel consumption, lowering it by 10.42% compared to actual driving scenarios.
arXiv Detail & Related papers (2024-07-22T16:48:37Z) - MetaFollower: Adaptable Personalized Autonomous Car Following [63.90050686330677]
We propose an adaptable personalized car-following framework - MetaFollower.
We first utilize Model-Agnostic Meta-Learning (MAML) to extract common driving knowledge from various CF events.
We additionally combine Long Short-Term Memory (LSTM) and Intelligent Driver Model (IDM) to reflect temporal heterogeneity with high interpretability.
arXiv Detail & Related papers (2024-06-23T15:30:40Z) - MOTO: Offline Pre-training to Online Fine-tuning for Model-based Robot
Learning [52.101643259906915]
We study the problem of offline pre-training and online fine-tuning for reinforcement learning from high-dimensional observations.
Existing model-based offline RL methods are not suitable for offline-to-online fine-tuning in high-dimensional domains.
We propose an on-policy model-based method that can efficiently reuse prior data through model-based value expansion and policy regularization.
arXiv Detail & Related papers (2024-01-06T21:04:31Z) - RACER: Rational Artificial Intelligence Car-following-model Enhanced by
Reality [51.244807332133696]
This paper introduces RACER, a cutting-edge deep learning car-following model to predict Adaptive Cruise Control (ACC) driving behavior.
Unlike conventional models, RACER effectively integrates Rational Driving Constraints (RDCs), crucial tenets of actual driving.
RACER excels across key metrics, such as acceleration, velocity, and spacing, registering zero violations.
arXiv Detail & Related papers (2023-12-12T06:21:30Z) - Data-efficient Deep Reinforcement Learning for Vehicle Trajectory
Control [6.144517901919656]
Reinforcement learning (RL) promises to achieve control performance superior to classical approaches.
Standard RL approaches like soft-actor critic (SAC) require extensive amounts of training data to be collected.
We apply recently developed data-efficient deep RL methods to vehicle trajectory control.
arXiv Detail & Related papers (2023-11-30T09:38:59Z) - Eco-Driving Control of Connected and Automated Vehicles using Neural
Network based Rollout [0.0]
Connected and autonomous vehicles have the potential to minimize energy consumption.
Existing deterministic and methods created to solve the eco-driving problem generally suffer from high computational and memory requirements.
This work proposes a hierarchical multi-horizon optimization framework implemented via a neural network.
arXiv Detail & Related papers (2023-10-16T23:13:51Z) - FastRLAP: A System for Learning High-Speed Driving via Deep RL and
Autonomous Practicing [71.76084256567599]
We present a system that enables an autonomous small-scale RC car to drive aggressively from visual observations using reinforcement learning (RL)
Our system, FastRLAP (faster lap), trains autonomously in the real world, without human interventions, and without requiring any simulation or expert demonstrations.
The resulting policies exhibit emergent aggressive driving skills, such as timing braking and acceleration around turns and avoiding areas which impede the robot's motion, approaching the performance of a human driver using a similar first-person interface over the course of training.
arXiv Detail & Related papers (2023-04-19T17:33:47Z) - Driver Assistance Eco-driving and Transmission Control with Deep
Reinforcement Learning [2.064612766965483]
In this paper, a model-free deep reinforcement learning (RL) control agent is proposed for active Eco-driving assistance.
It trades-off fuel consumption against other driver-accommodation objectives, and learns optimal traction torque and transmission shifting policies from experience.
It shows superior performance in minimizing fuel consumption compared to a baseline controller that has full knowledge of fuel-efficiency tables.
arXiv Detail & Related papers (2022-12-15T02:52:07Z) - An Energy Consumption Model for Electrical Vehicle Networks via Extended
Federated-learning [50.85048976506701]
This paper proposes a novel solution to range anxiety based on a federated-learning model.
It is capable of estimating battery consumption and providing energy-efficient route planning for vehicle networks.
arXiv Detail & Related papers (2021-11-13T15:03:44Z) - Data Freshness and Energy-Efficient UAV Navigation Optimization: A Deep
Reinforcement Learning Approach [88.45509934702913]
We design a navigation policy for multiple unmanned aerial vehicles (UAVs) where mobile base stations (BSs) are deployed.
We incorporate different contextual information such as energy and age of information (AoI) constraints to ensure the data freshness at the ground BS.
By applying the proposed trained model, an effective real-time trajectory policy for the UAV-BSs captures the observable network states over time.
arXiv Detail & Related papers (2020-02-21T07:29:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.