Constrained Optimal Fuel Consumption of HEV:Considering the Observational Perturbation
- URL: http://arxiv.org/abs/2410.20913v1
- Date: Mon, 28 Oct 2024 10:45:42 GMT
- Title: Constrained Optimal Fuel Consumption of HEV:Considering the Observational Perturbation
- Authors: Shuchang Yan, Haoran Sun,
- Abstract summary: We aim to minimize fuel consumption while maintaining SOC balance under observational perturbations in SOC and speed.
This work first worldwide uses seven training approaches to solve the COFC problem under five types of perturbations.
- Score: 12.936592572736908
- License:
- Abstract: We assume accurate observation of battery state of charge (SOC) and precise speed curves when addressing the constrained optimal fuel consumption (COFC) problem via constrained reinforcement learning (CRL). However, in practice, SOC measurements are often distorted by noise or confidentiality protocols, and actual reference speeds may deviate from expectations. We aim to minimize fuel consumption while maintaining SOC balance under observational perturbations in SOC and speed. This work first worldwide uses seven training approaches to solve the COFC problem under five types of perturbations, including one based on a uniform distribution, one designed to maximize rewards, one aimed at maximizing costs, and one along with its improved version that seeks to decrease reward on Toyota Hybrid Systems (THS) under New European Driving Cycle (NEDC) condition. The result verifies that the six can successfully solve the COFC problem under observational perturbations, and we further compare the robustness and safety of these training approaches and analyze their impact on optimal fuel consumption.
Related papers
- Constrained Optimal Fuel Consumption of HEV: A Constrained Reinforcement Learning Approach [0.0]
This work provides the mathematical expression of constrained optimal fuel consumption (COFC) from the perspective of constrained reinforcement learning (CRL)
Two mainstream approaches of CRL, constrained variational policy optimization (CVPO) and Lagrangian-based approaches, are utilized for the first time to obtain the vehicle's minimum fuel consumption.
arXiv Detail & Related papers (2024-03-12T10:42:32Z) - Catastrophic Overfitting: A Potential Blessing in Disguise [51.996943482875366]
Fast Adversarial Training (FAT) has gained increasing attention within the research community owing to its efficacy in improving adversarial robustness.
Although existing FAT approaches have made strides in mitigating CO, the ascent of adversarial robustness occurs with a non-negligible decline in classification accuracy on clean samples.
We employ the feature activation differences between clean and adversarial examples to analyze the underlying causes of CO.
We harness CO to achieve attack obfuscation', aiming to bolster model performance.
arXiv Detail & Related papers (2024-02-28T10:01:44Z) - COPR: Continual Human Preference Learning via Optimal Policy
Regularization [56.1193256819677]
Reinforcement Learning from Human Feedback (RLHF) is commonly utilized to improve the alignment of Large Language Models (LLMs) with human preferences.
We propose the Continual Optimal Policy Regularization (COPR) method, which draws inspiration from the optimal policy theory.
arXiv Detail & Related papers (2024-02-22T02:20:08Z) - EnduRL: Enhancing Safety, Stability, and Efficiency of Mixed Traffic Under Real-World Perturbations Via Reinforcement Learning [1.7273380623090846]
We analyze real-world driving trajectories and extract a wide range of acceleration profiles.
We then incorporates these profiles into simulations for training RVs to mitigate congestion.
Our RVs demonstrate significant improvements: safety by up to 66%, efficiency by up to 54%, and stability by up to 97%.
arXiv Detail & Related papers (2023-11-21T00:45:13Z) - Deep Reinforcement Learning-based Intelligent Traffic Signal Controls
with Optimized CO2 emissions [6.851243292023835]
Transportation networks face the challenge of sub-optimal control policies that can have adverse effects on human health, the environment, and contribute to traffic congestion.
Despite several adaptive traffic signal controllers in literature, limited research has been conducted on their comparative performance.
We propose EcoLight, a reward shaping scheme for reinforcement learning algorithms that not only reduces CO2 emissions but also achieves competitive results in metrics such as travel time.
arXiv Detail & Related papers (2023-10-19T19:54:47Z) - Hybrid Reinforcement Learning for Optimizing Pump Sustainability in
Real-World Water Distribution Networks [55.591662978280894]
This article addresses the pump-scheduling optimization problem to enhance real-time control of real-world water distribution networks (WDNs)
Our primary objectives are to adhere to physical operational constraints while reducing energy consumption and operational costs.
Traditional optimization techniques, such as evolution-based and genetic algorithms, often fall short due to their lack of convergence guarantees.
arXiv Detail & Related papers (2023-10-13T21:26:16Z) - Learning to Sail Dynamic Networks: The MARLIN Reinforcement Learning
Framework for Congestion Control in Tactical Environments [53.08686495706487]
This paper proposes an RL framework that leverages an accurate and parallelizable emulation environment to reenact the conditions of a tactical network.
We evaluate our RL learning framework by training a MARLIN agent in conditions replicating a bottleneck link transition between a Satellite Communication (SATCOM) and an UHF Wide Band (UHF) radio link.
arXiv Detail & Related papers (2023-06-27T16:15:15Z) - CCE: Sample Efficient Sparse Reward Policy Learning for Robotic Navigation via Confidence-Controlled Exploration [72.24964965882783]
Confidence-Controlled Exploration (CCE) is designed to enhance the training sample efficiency of reinforcement learning algorithms for sparse reward settings such as robot navigation.
CCE is based on a novel relationship we provide between gradient estimation and policy entropy.
We demonstrate through simulated and real-world experiments that CCE outperforms conventional methods that employ constant trajectory lengths and entropy regularization.
arXiv Detail & Related papers (2023-06-09T18:45:15Z) - Adaptive Frequency Green Light Optimal Speed Advisory based on Hybrid
Actor-Critic Reinforcement Learning [2.257737378757467]
GLOSA system suggests speeds to vehicles to assist them in passing through intersections during green intervals.
Previous research has focused on optimizing the GLOSA algorithm, neglecting the frequency of speed advisory.
We propose an Adaptive Frequency GLOSA model based on Hybrid Proximal Policy Optimization (H-PPO) method.
arXiv Detail & Related papers (2023-06-07T01:16:45Z) - Revealing the real-world CO2 emission reduction of ridesplitting and its
determinants based on machine learning [12.864925081071684]
This study calculates the CO2 emissions of shared rides (ridesplitting) and their substituted single rides (regular ridesourcing) to estimate the CO2 emission reduction of each ridesplitting trip.
The results show that not all ridesplitting trips reduce emissions from ridesourcing in the real world.
arXiv Detail & Related papers (2022-04-02T06:25:48Z) - Optimizing Mixed Autonomy Traffic Flow With Decentralized Autonomous
Vehicles and Multi-Agent RL [63.52264764099532]
We study the ability of autonomous vehicles to improve the throughput of a bottleneck using a fully decentralized control scheme in a mixed autonomy setting.
We apply multi-agent reinforcement algorithms to this problem and demonstrate that significant improvements in bottleneck throughput, from 20% at a 5% penetration rate to 33% at a 40% penetration rate, can be achieved.
arXiv Detail & Related papers (2020-10-30T22:06:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.