Safe Neural Control for Non-Affine Control Systems with Differentiable
Control Barrier Functions
- URL: http://arxiv.org/abs/2309.04492v1
- Date: Wed, 6 Sep 2023 05:35:48 GMT
- Title: Safe Neural Control for Non-Affine Control Systems with Differentiable
Control Barrier Functions
- Authors: Wei Xiao and Ross Allen and Daniela Rus
- Abstract summary: This paper addresses the problem of safety-critical control for non-affine control systems.
It has been shown that optimizing quadratic costs subject to state and control constraints can be sub-optimally reduced to a sequence of quadratic programs (QPs) by using Control Barrier Functions (CBFs)
We incorporate higher-order CBFs into neural ordinary differential equation-based learning models as differentiable CBFs to guarantee safety for non-affine control systems.
- Score: 58.19198103790931
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper addresses the problem of safety-critical control for non-affine
control systems. It has been shown that optimizing quadratic costs subject to
state and control constraints can be sub-optimally reduced to a sequence of
quadratic programs (QPs) by using Control Barrier Functions (CBFs). Our
recently proposed High Order CBFs (HOCBFs) can accommodate constraints of
arbitrary relative degree. The main challenges in this approach are that it
requires affine control dynamics and the solution of the CBF-based QP is
sub-optimal since it is solved point-wise. To address these challenges, we
incorporate higher-order CBFs into neural ordinary differential equation-based
learning models as differentiable CBFs to guarantee safety for non-affine
control systems. The differentiable CBFs are trainable in terms of their
parameters, and thus, they can address the conservativeness of CBFs such that
the system state will not stay unnecessarily far away from safe set boundaries.
Moreover, the imitation learning model is capable of learning complex and
optimal control policies that are usually intractable online. We illustrate the
effectiveness of the proposed framework on LiDAR-based autonomous driving and
compare it with existing methods.
Related papers
- Domain Adaptive Safety Filters via Deep Operator Learning [5.62479170374811]
We propose a self-supervised deep operator learning framework that learns the mapping from environmental parameters to the corresponding CBF.
We demonstrate the effectiveness of the method through numerical experiments on navigation tasks involving dynamic obstacles.
arXiv Detail & Related papers (2024-10-18T15:10:55Z) - Pareto Control Barrier Function for Inner Safe Set Maximization Under Input Constraints [50.920465513162334]
We introduce the PCBF algorithm to maximize the inner safe set of dynamical systems under input constraints.
We validate its effectiveness through comparison with Hamilton-Jacobi reachability for an inverted pendulum and through simulations on a 12-dimensional quadrotor system.
Results show that the PCBF consistently outperforms existing methods, yielding larger safe sets and ensuring safety under input constraints.
arXiv Detail & Related papers (2024-10-05T18:45:19Z) - Reinforcement Learning-based Receding Horizon Control using Adaptive Control Barrier Functions for Safety-Critical Systems [14.166970599802324]
Optimal control methods provide solutions to safety-critical problems but easily become intractable.
We propose a Reinforcement Learning-based Receding Horizon Control approach leveraging Model Predictive Control.
We validate our method by applying it to the challenging automated merging control problem for Connected and Automated Vehicles.
arXiv Detail & Related papers (2024-03-26T02:49:08Z) - Learning Performance-Oriented Control Barrier Functions Under Complex Safety Constraints and Limited Actuation [5.62479170374811]
Control Barrier Functions (CBFs) provide an elegant framework for constraining nonlinear control system dynamics.
We introduce a novel self-supervised learning framework to comprehensively address these challenges.
We validate our approach on a 2D double integrator (DI) system and a 7D fixed-wing aircraft system.
arXiv Detail & Related papers (2024-01-11T02:51:49Z) - Learning Feasibility Constraints for Control Barrier Functions [8.264868845642843]
We employ machine learning techniques to ensure the feasibility of Quadratic Programs (QPs)
We propose a sampling-based learning approach to learn a new feasibility constraint for CBFs.
We demonstrate the advantages of the proposed learning approach to constrained optimal control problems.
arXiv Detail & Related papers (2023-03-10T16:29:20Z) - Learning Robust Output Control Barrier Functions from Safe Expert Demonstrations [50.37808220291108]
This paper addresses learning safe output feedback control laws from partial observations of expert demonstrations.
We first propose robust output control barrier functions (ROCBFs) as a means to guarantee safety.
We then formulate an optimization problem to learn ROCBFs from expert demonstrations that exhibit safe system behavior.
arXiv Detail & Related papers (2021-11-18T23:21:00Z) - Pointwise Feasibility of Gaussian Process-based Safety-Critical Control
under Model Uncertainty [77.18483084440182]
Control Barrier Functions (CBFs) and Control Lyapunov Functions (CLFs) are popular tools for enforcing safety and stability of a controlled system, respectively.
We present a Gaussian Process (GP)-based approach to tackle the problem of model uncertainty in safety-critical controllers that use CBFs and CLFs.
arXiv Detail & Related papers (2021-06-13T23:08:49Z) - Reinforcement Learning for Safety-Critical Control under Model
Uncertainty, using Control Lyapunov Functions and Control Barrier Functions [96.63967125746747]
Reinforcement learning framework learns the model uncertainty present in the CBF and CLF constraints.
RL-CBF-CLF-QP addresses the problem of model uncertainty in the safety constraints.
arXiv Detail & Related papers (2020-04-16T10:51:33Z) - Learning Control Barrier Functions from Expert Demonstrations [69.23675822701357]
We propose a learning based approach to safe controller synthesis based on control barrier functions (CBFs)
We analyze an optimization-based approach to learning a CBF that enjoys provable safety guarantees under suitable Lipschitz assumptions on the underlying dynamical system.
To the best of our knowledge, these are the first results that learn provably safe control barrier functions from data.
arXiv Detail & Related papers (2020-04-07T12:29:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.