An Empirical Analysis of the Use of Real-Time Reachability for the
Safety Assurance of Autonomous Vehicles
- URL: http://arxiv.org/abs/2205.01419v1
- Date: Tue, 3 May 2022 11:12:29 GMT
- Title: An Empirical Analysis of the Use of Real-Time Reachability for the
Safety Assurance of Autonomous Vehicles
- Authors: Patrick Musau, Nathaniel Hamilton, Diego Manzanas Lopez, Preston
Robinette, Taylor T. Johnson
- Abstract summary: We propose using a real-time reachability algorithm for the implementation of the simplex architecture to assure the safety of a 1/10 scale open source autonomous vehicle platform.
In our approach, the need to analyze an underlying controller is abstracted away, instead focusing on the effects of the controller's decisions on the system's future states.
- Score: 7.1169864450668845
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Recent advances in machine learning technologies and sensing have paved the
way for the belief that safe, accessible, and convenient autonomous vehicles
may be realized in the near future. Despite tremendous advances within this
context, fundamental challenges around safety and reliability are limiting
their arrival and comprehensive adoption. Autonomous vehicles are often tasked
with operating in dynamic and uncertain environments. As a result, they often
make use of highly complex components, such as machine learning approaches, to
handle the nuances of sensing, actuation, and control. While these methods are
highly effective, they are notoriously difficult to assure. Moreover, within
uncertain and dynamic environments, design time assurance analyses may not be
sufficient to guarantee safety. Thus, it is critical to monitor the correctness
of these systems at runtime. One approach for providing runtime assurance of
systems with components that may not be amenable to formal analysis is the
simplex architecture, where an unverified component is wrapped with a safety
controller and a switching logic designed to prevent dangerous behavior. In
this paper, we propose using a real-time reachability algorithm for the
implementation of the simplex architecture to assure the safety of a 1/10 scale
open source autonomous vehicle platform known as F1/10. The reachability
algorithm that we leverage (a) provides provable guarantees of safety, and (b)
is used to detect potentially unsafe scenarios. In our approach, the need to
analyze an underlying controller is abstracted away, instead focusing on the
effects of the controller's decisions on the system's future states. We
demonstrate the efficacy of our architecture through a vast set of experiments
conducted both in simulation and on an embedded hardware platform.
Related papers
- Handling Long-Term Safety and Uncertainty in Safe Reinforcement Learning [17.856459823003277]
Safety is one of the key issues preventing the deployment of reinforcement learning techniques in real-world robots.
In this paper, we bridge the gap by extending the safe exploration method, ATACOM, with learnable constraints.
arXiv Detail & Related papers (2024-09-18T15:08:41Z) - Searching for Optimal Runtime Assurance via Reachability and
Reinforcement Learning [2.422636931175853]
runtime assurance system (RTA) for a given plant enables the exercise of an untrusted or experimental controller while assuring safety with a backup controller.
Existing RTA design strategies are well-known to be overly conservative and, in principle, can lead to safety violations.
In this paper, we formulate the optimal RTA design problem and present a new approach for solving it.
arXiv Detail & Related papers (2023-10-06T14:45:57Z) - Simulation-based Safety Assurance for an AVP System incorporating
Learning-Enabled Components [0.6526824510982802]
Testing, verification and validation AD/ADAS safety-critical applications remain as one the main challenges.
We explain the simulation-based development platform that is designed to verify and validate safety-critical learning-enabled systems.
arXiv Detail & Related papers (2023-09-28T09:00:31Z) - Leveraging Traceability to Integrate Safety Analysis Artifacts into the
Software Development Process [51.42800587382228]
Safety assurance cases (SACs) can be challenging to maintain during system evolution.
We propose a solution that leverages software traceability to connect relevant system artifacts to safety analysis models.
We elicit design rationales for system changes to help safety stakeholders analyze the impact of system changes on safety.
arXiv Detail & Related papers (2023-07-14T16:03:27Z) - Evaluating Model-free Reinforcement Learning toward Safety-critical
Tasks [70.76757529955577]
This paper revisits prior work in this scope from the perspective of state-wise safe RL.
We propose Unrolling Safety Layer (USL), a joint method that combines safety optimization and safety projection.
To facilitate further research in this area, we reproduce related algorithms in a unified pipeline and incorporate them into SafeRL-Kit.
arXiv Detail & Related papers (2022-12-12T06:30:17Z) - Recursively Feasible Probabilistic Safe Online Learning with Control Barrier Functions [60.26921219698514]
We introduce a model-uncertainty-aware reformulation of CBF-based safety-critical controllers.
We then present the pointwise feasibility conditions of the resulting safety controller.
We use these conditions to devise an event-triggered online data collection strategy.
arXiv Detail & Related papers (2022-08-23T05:02:09Z) - Differentiable Control Barrier Functions for Vision-based End-to-End
Autonomous Driving [100.57791628642624]
We introduce a safety guaranteed learning framework for vision-based end-to-end autonomous driving.
We design a learning system equipped with differentiable control barrier functions (dCBFs) that is trained end-to-end by gradient descent.
arXiv Detail & Related papers (2022-03-04T16:14:33Z) - Closing the Closed-Loop Distribution Shift in Safe Imitation Learning [80.05727171757454]
We treat safe optimization-based control strategies as experts in an imitation learning problem.
We train a learned policy that can be cheaply evaluated at run-time and that provably satisfies the same safety guarantees as the expert.
arXiv Detail & Related papers (2021-02-18T05:11:41Z) - Scalable Learning of Safety Guarantees for Autonomous Systems using
Hamilton-Jacobi Reachability [18.464688553299663]
Methods like Hamilton-Jacobi reachability can provide guaranteed safe sets and controllers for such systems.
As the system is operating, it may learn new knowledge about these uncertainties and should therefore update its safety analysis accordingly.
In this paper we synthesize several techniques to speed up computation: decomposition, warm-starting, and adaptive grids.
arXiv Detail & Related papers (2021-01-15T00:13:01Z) - Dos and Don'ts of Machine Learning in Computer Security [74.1816306998445]
Despite great potential, machine learning in security is prone to subtle pitfalls that undermine its performance.
We identify common pitfalls in the design, implementation, and evaluation of learning-based security systems.
We propose actionable recommendations to support researchers in avoiding or mitigating the pitfalls where possible.
arXiv Detail & Related papers (2020-10-19T13:09:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.