DRIP: Domain Refinement Iteration with Polytopes for Backward
Reachability Analysis of Neural Feedback Loops
- URL: http://arxiv.org/abs/2212.04646v1
- Date: Fri, 9 Dec 2022 03:06:58 GMT
- Title: DRIP: Domain Refinement Iteration with Polytopes for Backward
Reachability Analysis of Neural Feedback Loops
- Authors: Michael Everett, Rudy Bunel, Shayegan Omidshafiei
- Abstract summary: This work investigates backward reachability as a framework for providing collision avoidance guarantees for systems controlled by neural network (NN) policies.
Because NNs are typically not invertible, existing methods assume a domain over which to relax the NN, which causes loose over-approximations of the set of states.
We introduce DRIP, an algorithm with a refinement loop on the relaxation domain, which substantially tightens the BP set bounds.
- Score: 12.706980346861986
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Safety certification of data-driven control techniques remains a major open
problem. This work investigates backward reachability as a framework for
providing collision avoidance guarantees for systems controlled by neural
network (NN) policies. Because NNs are typically not invertible, existing
methods conservatively assume a domain over which to relax the NN, which causes
loose over-approximations of the set of states that could lead the system into
the obstacle (i.e., backprojection (BP) sets). To address this issue, we
introduce DRIP, an algorithm with a refinement loop on the relaxation domain,
which substantially tightens the BP set bounds. Furthermore, we introduce a
formulation that enables directly obtaining closed-form representations of
polytopes to bound the BP sets tighter than prior work, which required solving
linear programs and using hyper-rectangles. Furthermore, this work extends the
NN relaxation algorithm to handle polytope domains, which further tightens the
bounds on BP sets. DRIP is demonstrated in numerical experiments on control
systems, including a ground robot controlled by a learned NN obstacle avoidance
policy.
Related papers
- Probabilistic Reach-Avoid for Bayesian Neural Networks [71.67052234622781]
We show that an optimal synthesis algorithm can provide more than a four-fold increase in the number of certifiable states.
The algorithm is able to provide more than a three-fold increase in the average guaranteed reach-avoid probability.
arXiv Detail & Related papers (2023-10-03T10:52:21Z) - Safe Neural Control for Non-Affine Control Systems with Differentiable
Control Barrier Functions [58.19198103790931]
This paper addresses the problem of safety-critical control for non-affine control systems.
It has been shown that optimizing quadratic costs subject to state and control constraints can be sub-optimally reduced to a sequence of quadratic programs (QPs) by using Control Barrier Functions (CBFs)
We incorporate higher-order CBFs into neural ordinary differential equation-based learning models as differentiable CBFs to guarantee safety for non-affine control systems.
arXiv Detail & Related papers (2023-09-06T05:35:48Z) - Backward Reachability Analysis of Neural Feedback Loops: Techniques for
Linear and Nonlinear Systems [59.57462129637796]
This paper presents a backward reachability approach for safety verification of closed-loop systems with neural networks (NNs)
The presence of NNs in the feedback loop presents a unique set of problems due to the nonlinearities in their activation functions and because NN models are generally not invertible.
We present frameworks for calculating BP over-approximations for both linear and nonlinear systems with control policies represented by feedforward NNs.
arXiv Detail & Related papers (2022-09-28T13:17:28Z) - Backward Reachability Analysis for Neural Feedback Loops [40.989393438716476]
This paper presents a backward reachability approach for safety verification of closed-loop systems with neural networks (NNs)
The presence of NNs in the feedback loop presents a unique set of problems due to the nonlinearities in their activation functions and because NN models are generally not invertible.
We present an algorithm to iteratively find BP set estimates over a given time horizon and demonstrate the ability to reduce conservativeness by up to 88% with low additional computational cost.
arXiv Detail & Related papers (2022-04-14T01:13:14Z) - Distributed neural network control with dependability guarantees: a
compositional port-Hamiltonian approach [0.0]
Large-scale cyber-physical systems require that control policies are distributed, that is, that they only rely on local real-time measurements and communication with neighboring agents.
Recent work has proposed training Neural Network (NN) distributed controllers.
A main challenge of NN controllers is that they are not dependable during and after training, that is, the closed-loop system may be unstable, and the training may fail due to vanishing and exploding gradients.
arXiv Detail & Related papers (2021-12-16T17:37:11Z) - OVERT: An Algorithm for Safety Verification of Neural Network Control
Policies for Nonlinear Systems [31.3812947670948]
We present OVERT: a sound algorithm for safety verification of neural network control policies.
The central concept of OVERT is to abstract nonlinear functions with a set of optimally tight piecewise linear bounds.
Overt compares favorably to existing methods both in time and in tightness of the reachable set.
arXiv Detail & Related papers (2021-08-03T00:41:27Z) - Shortest-Path Constrained Reinforcement Learning for Sparse Reward Tasks [59.419152768018506]
We show that any optimal policy necessarily satisfies the k-SP constraint.
We propose a novel cost function that penalizes the policy violating SP constraint, instead of completely excluding it.
Our experiments on MiniGrid, DeepMind Lab, Atari, and Fetch show that the proposed method significantly improves proximal policy optimization (PPO)
arXiv Detail & Related papers (2021-07-13T21:39:21Z) - Certification of Iterative Predictions in Bayesian Neural Networks [79.15007746660211]
We compute lower bounds for the probability that trajectories of the BNN model reach a given set of states while avoiding a set of unsafe states.
We use the lower bounds in the context of control and reinforcement learning to provide safety certification for given control policies.
arXiv Detail & Related papers (2021-05-21T05:23:57Z) - Chance-Constrained Control with Lexicographic Deep Reinforcement
Learning [77.34726150561087]
This paper proposes a lexicographic Deep Reinforcement Learning (DeepRL)-based approach to chance-constrained Markov Decision Processes.
A lexicographic version of the well-known DeepRL algorithm DQN is also proposed and validated via simulations.
arXiv Detail & Related papers (2020-10-19T13:09:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.