Self-Awareness Safety of Deep Reinforcement Learning in Road Traffic
Junction Driving
- URL: http://arxiv.org/abs/2201.08116v1
- Date: Thu, 20 Jan 2022 11:21:33 GMT
- Title: Self-Awareness Safety of Deep Reinforcement Learning in Road Traffic
Junction Driving
- Authors: Zehong Cao, Jie Yun
- Abstract summary: In a road traffic junction scenario, the vehicle typically receives partial observations from the transportation environment.
In this study, we evaluated the safety performance of three baseline DRL models (DQN, A2C, and PPO)
Our proposed self-awareness attention-DQN can significantly improve the safety performance in intersection and roundabout scenarios.
- Score: 20.85562165500152
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Autonomous driving has been at the forefront of public interest, and a
pivotal debate to widespread concerns is safety in the transportation system.
Deep reinforcement learning (DRL) has been applied to autonomous driving to
provide solutions for obstacle avoidance. However, in a road traffic junction
scenario, the vehicle typically receives partial observations from the
transportation environment, while DRL needs to rely on long-term rewards to
train a reliable model by maximising the cumulative rewards, which may take the
risk when exploring new actions and returning either a positive reward or a
penalty in the case of collisions. Although safety concerns are usually
considered in the design of a reward function, they are not fully considered as
the critical metric to directly evaluate the effectiveness of DRL algorithms in
autonomous driving. In this study, we evaluated the safety performance of three
baseline DRL models (DQN, A2C, and PPO) and proposed a self-awareness module
from an attention mechanism for DRL to improve the safety evaluation for an
anomalous vehicle in a complex road traffic junction environment, such as
intersection and roundabout scenarios, based on four metrics: collision rate,
success rate, freezing rate, and total reward. Our two experimental results in
the training and testing phases revealed the baseline DRL with poor safety
performance, while our proposed self-awareness attention-DQN can significantly
improve the safety performance in intersection and roundabout scenarios.
Related papers
- Autonomous Vehicle Decision-Making Framework for Considering Malicious
Behavior at Unsignalized Intersections [7.245712580297489]
In Autonomous Vehicles, reward signals are set as regular rewards regarding feedback factors such as safety and efficiency.
In this paper, safety gains are modulated by variable weighting parameters to ensure that safety can be emphasized more in emergency situations.
The decision framework enables Autonomous Vehicles to make informed decisions when encountering vehicles with potentially malicious behaviors at unsignalized intersections.
arXiv Detail & Related papers (2024-09-11T03:57:44Z) - SECRM-2D: RL-Based Efficient and Comfortable Route-Following Autonomous Driving with Analytic Safety Guarantees [5.156059061769101]
SECRM-2D is an RL autonomous driving controller that balances optimization of efficiency and comfort and follows a fixed route.
We evaluate SECRM-2D against several learning and non-learning baselines in simulated test scenarios.
arXiv Detail & Related papers (2024-07-23T21:54:39Z) - RACER: Epistemic Risk-Sensitive RL Enables Fast Driving with Fewer Crashes [57.319845580050924]
We propose a reinforcement learning framework that combines risk-sensitive control with an adaptive action space curriculum.
We show that our algorithm is capable of learning high-speed policies for a real-world off-road driving task.
arXiv Detail & Related papers (2024-05-07T23:32:36Z) - Uniformly Safe RL with Objective Suppression for Multi-Constraint Safety-Critical Applications [73.58451824894568]
The widely adopted CMDP model constrains the risks in expectation, which makes room for dangerous behaviors in long-tail states.
In safety-critical domains, such behaviors could lead to disastrous outcomes.
We propose Objective Suppression, a novel method that adaptively suppresses the task reward maximizing objectives according to a safety critic.
arXiv Detail & Related papers (2024-02-23T23:22:06Z) - Safety-aware Causal Representation for Trustworthy Offline Reinforcement
Learning in Autonomous Driving [33.672722472758636]
offline Reinforcement Learning(RL) approaches exhibit notable efficacy in addressing sequential decision-making problems from offline datasets.
We introduce the saFety-aware strUctured Scenario representatION ( Fusion) to facilitate the learning of a generalizable end-to-end driving policy.
Empirical evidence in various driving scenarios attests that Fusion significantly enhances the safety and generalizability of autonomous driving agents.
arXiv Detail & Related papers (2023-10-31T18:21:24Z) - CAT: Closed-loop Adversarial Training for Safe End-to-End Driving [54.60865656161679]
Adversarial Training (CAT) is a framework for safe end-to-end driving in autonomous vehicles.
Cat aims to continuously improve the safety of driving agents by training the agent on safety-critical scenarios.
Cat can effectively generate adversarial scenarios countering the agent being trained.
arXiv Detail & Related papers (2023-10-19T02:49:31Z) - Towards Safe Autonomous Driving Policies using a Neuro-Symbolic Deep
Reinforcement Learning Approach [6.961253535504979]
This paper introduces a novel neuro-symbolic model-free DRL approach, called DRL with Symbolic Logics (DRLSL)
It combines the strengths of DRL (learning from experience) and symbolic first-order logics (knowledge-driven reasoning) to enable safe learning in real-time interactions of autonomous driving within real environments.
We have implemented the DRLSL framework in autonomous driving using the highD dataset and demonstrated that our method successfully avoids unsafe actions during both the training and testing phases.
arXiv Detail & Related papers (2023-07-03T19:43:21Z) - A Multiplicative Value Function for Safe and Efficient Reinforcement
Learning [131.96501469927733]
We propose a safe model-free RL algorithm with a novel multiplicative value function consisting of a safety critic and a reward critic.
The safety critic predicts the probability of constraint violation and discounts the reward critic that only estimates constraint-free returns.
We evaluate our method in four safety-focused environments, including classical RL benchmarks augmented with safety constraints and robot navigation tasks with images and raw Lidar scans as observations.
arXiv Detail & Related papers (2023-03-07T18:29:15Z) - Safe Reinforcement Learning for an Energy-Efficient Driver Assistance
System [1.8899300124593645]
Reinforcement learning (RL)-based driver assistance systems seek to improve fuel consumption via continual improvement of powertrain control actions.
In this paper, an exponential control barrier function (ECBF) is derived and utilized to filter unsafe actions proposed by an RL-based driver assistance system.
The proposed safe-RL scheme is trained and evaluated in car following scenarios where it is shown that it effectively avoids collision both during training and evaluation.
arXiv Detail & Related papers (2023-01-03T00:25:00Z) - Benchmarking Safe Deep Reinforcement Learning in Aquatic Navigation [78.17108227614928]
We propose a benchmark environment for Safe Reinforcement Learning focusing on aquatic navigation.
We consider a value-based and policy-gradient Deep Reinforcement Learning (DRL)
We also propose a verification strategy that checks the behavior of the trained models over a set of desired properties.
arXiv Detail & Related papers (2021-12-16T16:53:56Z) - Cautious Adaptation For Reinforcement Learning in Safety-Critical
Settings [129.80279257258098]
Reinforcement learning (RL) in real-world safety-critical target settings like urban driving is hazardous.
We propose a "safety-critical adaptation" task setting: an agent first trains in non-safety-critical "source" environments.
We propose a solution approach, CARL, that builds on the intuition that prior experience in diverse environments equips an agent to estimate risk.
arXiv Detail & Related papers (2020-08-15T01:40:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.