Safety Certification for Stochastic Systems via Neural Barrier Functions
- URL: http://arxiv.org/abs/2206.01463v1
- Date: Fri, 3 Jun 2022 09:06:02 GMT
- Title: Safety Certification for Stochastic Systems via Neural Barrier Functions
- Authors: Frederik Baymler Mathiesen, Simeon Calvert, Luca Laurenti
- Abstract summary: barrier functions can be used to provide non-trivial certificates of safety for non-linear systems.
We parameterize a barrier function as a neural network and show that robust training of neural networks can be successfully employed to find barrier functions.
We show that our approach outperforms existing methods in several case studies and often returns certificates of safety that are orders of magnitude larger.
- Score: 3.7491936479803054
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Providing non-trivial certificates of safety for non-linear stochastic
systems is an important open problem that limits the wider adoption of
autonomous systems in safety-critical applications. One promising solution to
address this problem is barrier functions. The composition of a barrier
function with a stochastic system forms a supermartingale, thus enabling the
computation of the probability that the system stays in a safe set over a
finite time horizon via martingale inequalities. However, existing approaches
to find barrier functions for stochastic systems generally rely on convex
optimization programs that restrict the search of a barrier to a small class of
functions such as low degree SoS polynomials and can be computationally
expensive. In this paper, we parameterize a barrier function as a neural
network and show that techniques for robust training of neural networks can be
successfully employed to find neural barrier functions. Specifically, we
leverage bound propagation techniques to certify that a neural network
satisfies the conditions to be a barrier function via linear programming and
then employ the resulting bounds at training time to enforce the satisfaction
of these conditions. We also present a branch-and-bound scheme that makes the
certification framework scalable. We show that our approach outperforms
existing methods in several case studies and often returns certificates of
safety that are orders of magnitude larger.
Related papers
- Learning Controlled Stochastic Differential Equations [61.82896036131116]
This work proposes a novel method for estimating both drift and diffusion coefficients of continuous, multidimensional, nonlinear controlled differential equations with non-uniform diffusion.
We provide strong theoretical guarantees, including finite-sample bounds for (L2), (Linfty), and risk metrics, with learning rates adaptive to coefficients' regularity.
Our method is available as an open-source Python library.
arXiv Detail & Related papers (2024-11-04T11:09:58Z) - Verification-Aided Learning of Neural Network Barrier Functions with
Termination Guarantees [6.9060054915724]
Barrier functions are a general framework for establishing a safety guarantee for a system.
There is no general method for finding these functions.
Recent approaches use self-supervised learning techniques to learn these functions.
arXiv Detail & Related papers (2024-03-12T04:29:43Z) - Robust Stochastically-Descending Unrolled Networks [85.6993263983062]
Deep unrolling is an emerging learning-to-optimize method that unrolls a truncated iterative algorithm in the layers of a trainable neural network.
We show that convergence guarantees and generalizability of the unrolled networks are still open theoretical problems.
We numerically assess unrolled architectures trained under the proposed constraints in two different applications.
arXiv Detail & Related papers (2023-12-25T18:51:23Z) - Enforcing Hard Constraints with Soft Barriers: Safe Reinforcement
Learning in Unknown Stochastic Environments [84.3830478851369]
We propose a safe reinforcement learning approach that can jointly learn the environment and optimize the control policy.
Our approach can effectively enforce hard safety constraints and significantly outperform CMDP-based baseline methods in system safe rate measured via simulations.
arXiv Detail & Related papers (2022-09-29T20:49:25Z) - Recursively Feasible Probabilistic Safe Online Learning with Control Barrier Functions [60.26921219698514]
We introduce a model-uncertainty-aware reformulation of CBF-based safety-critical controllers.
We then present the pointwise feasibility conditions of the resulting safety controller.
We use these conditions to devise an event-triggered online data collection strategy.
arXiv Detail & Related papers (2022-08-23T05:02:09Z) - Log Barriers for Safe Black-box Optimization with Application to Safe
Reinforcement Learning [72.97229770329214]
We introduce a general approach for seeking high dimensional non-linear optimization problems in which maintaining safety during learning is crucial.
Our approach called LBSGD is based on applying a logarithmic barrier approximation with a carefully chosen step size.
We demonstrate the effectiveness of our approach on minimizing violation in policy tasks in safe reinforcement learning.
arXiv Detail & Related papers (2022-07-21T11:14:47Z) - Stability Verification in Stochastic Control Systems via Neural Network
Supermartingales [17.558766911646263]
We present an approach for general nonlinear control problems with two novel aspects.
We use ranking supergales (RSMs) to certify a.s.asymptotic stability, and we present a method for learning neural networks.
arXiv Detail & Related papers (2021-12-17T13:05:14Z) - Constrained Feedforward Neural Network Training via Reachability
Analysis [0.0]
It remains an open challenge to train a neural network to obey safety constraints.
This work proposes a constrained method to simultaneously train and verify a feedforward neural network with rectified linear unit (ReLU) nonlinearities.
arXiv Detail & Related papers (2021-07-16T04:03:01Z) - Reach-SDP: Reachability Analysis of Closed-Loop Systems with Neural
Network Controllers via Semidefinite Programming [19.51345816555571]
We propose a novel forward reachability analysis method for the safety verification of linear time-varying systems with neural networks in feedback.
We show that we can compute these approximate reachable sets using semidefinite programming.
We illustrate our method in a quadrotor example, in which we first approximate a nonlinear model predictive controller via a deep neural network and then apply our analysis tool to certify finite-time reachability and constraint satisfaction of the closed-loop system.
arXiv Detail & Related papers (2020-04-16T18:48:25Z) - Learning Control Barrier Functions from Expert Demonstrations [69.23675822701357]
We propose a learning based approach to safe controller synthesis based on control barrier functions (CBFs)
We analyze an optimization-based approach to learning a CBF that enjoys provable safety guarantees under suitable Lipschitz assumptions on the underlying dynamical system.
To the best of our knowledge, these are the first results that learn provably safe control barrier functions from data.
arXiv Detail & Related papers (2020-04-07T12:29:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.