Verification-Aided Learning of Neural Network Barrier Functions with
Termination Guarantees
- URL: http://arxiv.org/abs/2403.07308v1
- Date: Tue, 12 Mar 2024 04:29:43 GMT
- Title: Verification-Aided Learning of Neural Network Barrier Functions with
Termination Guarantees
- Authors: Shaoru Chen, Lekan Molu, Mahyar Fazlyab
- Abstract summary: Barrier functions are a general framework for establishing a safety guarantee for a system.
There is no general method for finding these functions.
Recent approaches use self-supervised learning techniques to learn these functions.
- Score: 6.9060054915724
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Barrier functions are a general framework for establishing a safety guarantee
for a system. However, there is no general method for finding these functions.
To address this shortcoming, recent approaches use self-supervised learning
techniques to learn these functions using training data that are periodically
generated by a verification procedure, leading to a verification-aided learning
framework. Despite its immense potential in automating barrier function
synthesis, the verification-aided learning framework does not have termination
guarantees and may suffer from a low success rate of finding a valid barrier
function in practice. In this paper, we propose a holistic approach to address
these drawbacks. With a convex formulation of the barrier function synthesis,
we propose to first learn an empirically well-behaved NN basis function and
then apply a fine-tuning algorithm that exploits the convexity and
counterexamples from the verification failure to find a valid barrier function
with finite-step termination guarantees: if there exist valid barrier
functions, the fine-tuning algorithm is guaranteed to find one in a finite
number of iterations. We demonstrate that our fine-tuning method can
significantly boost the performance of the verification-aided learning
framework on examples of different scales and using various neural network
verifiers.
Related papers
- Benefits of Monotonicity in Safe Exploration with Gaussian Processes [50.71125084216603]
We consider the problem of sequentially maximising an unknown function over a set of actions.
We show that textscsffamily M-SafeUCB enjoys theoretical guarantees in terms of safety, a suitably-defined regret notion, and approximately finding the entire safe boundary.
arXiv Detail & Related papers (2022-11-03T02:52:30Z) - Offline Reinforcement Learning with Differentiable Function
Approximation is Provably Efficient [65.08966446962845]
offline reinforcement learning, which aims at optimizing decision-making strategies with historical data, has been extensively applied in real-life applications.
We take a step by considering offline reinforcement learning with differentiable function class approximation (DFA)
Most importantly, we show offline differentiable function approximation is provably efficient by analyzing the pessimistic fitted Q-learning algorithm.
arXiv Detail & Related papers (2022-10-03T07:59:42Z) - Log Barriers for Safe Black-box Optimization with Application to Safe
Reinforcement Learning [72.97229770329214]
We introduce a general approach for seeking high dimensional non-linear optimization problems in which maintaining safety during learning is crucial.
Our approach called LBSGD is based on applying a logarithmic barrier approximation with a carefully chosen step size.
We demonstrate the effectiveness of our approach on minimizing violation in policy tasks in safe reinforcement learning.
arXiv Detail & Related papers (2022-07-21T11:14:47Z) - Safety Certification for Stochastic Systems via Neural Barrier Functions [3.7491936479803054]
barrier functions can be used to provide non-trivial certificates of safety for non-linear systems.
We parameterize a barrier function as a neural network and show that robust training of neural networks can be successfully employed to find barrier functions.
We show that our approach outperforms existing methods in several case studies and often returns certificates of safety that are orders of magnitude larger.
arXiv Detail & Related papers (2022-06-03T09:06:02Z) - Stabilizing Q-learning with Linear Architectures for Provably Efficient
Learning [53.17258888552998]
This work proposes an exploration variant of the basic $Q$-learning protocol with linear function approximation.
We show that the performance of the algorithm degrades very gracefully under a novel and more permissive notion of approximation error.
arXiv Detail & Related papers (2022-06-01T23:26:51Z) - Safety and Liveness Guarantees through Reach-Avoid Reinforcement
Learning [24.56889192688925]
Reach-avoid optimal control problems are central to safety and liveness assurance for autonomous robotic systems.
Recent successes in reinforcement learning methods to approximately solve optimal control problems with performance objectives make their application to certification problems attractive.
Recent work has shown promise in extending the reinforcement learning machinery to handle safety-type problems, whose objective is not a sum, but a minimum (or maximum) over time.
arXiv Detail & Related papers (2021-12-23T00:44:38Z) - Part-X: A Family of Stochastic Algorithms for Search-Based Test
Generation with Probabilistic Guarantees [3.9119084077397863]
falsification has proven to be a practical and effective method for discovering erroneous behaviors in Cyber-Physical Systems.
Despite the constant improvements on the performance and applicability of falsification methods, they all share a common characteristic.
They are best-effort methods which do not provide any guarantees on the absence of erroneous behaviors (falsifiers) when the testing budget is exhausted.
arXiv Detail & Related papers (2021-10-20T19:05:00Z) - A Boosting Approach to Reinforcement Learning [59.46285581748018]
We study efficient algorithms for reinforcement learning in decision processes whose complexity is independent of the number of states.
We give an efficient algorithm that is capable of improving the accuracy of such weak learning methods.
arXiv Detail & Related papers (2021-08-22T16:00:45Z) - Learning Control Barrier Functions from Expert Demonstrations [69.23675822701357]
We propose a learning based approach to safe controller synthesis based on control barrier functions (CBFs)
We analyze an optimization-based approach to learning a CBF that enjoys provable safety guarantees under suitable Lipschitz assumptions on the underlying dynamical system.
To the best of our knowledge, these are the first results that learn provably safe control barrier functions from data.
arXiv Detail & Related papers (2020-04-07T12:29:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.