Learning Local Control Barrier Functions for Safety Control of Hybrid
Systems
- URL: http://arxiv.org/abs/2401.14907v1
- Date: Fri, 26 Jan 2024 14:38:43 GMT
- Title: Learning Local Control Barrier Functions for Safety Control of Hybrid
Systems
- Authors: Shuo Yang, Yu Chen, Xiang Yin, Rahul Mangharam
- Abstract summary: Safety is a primary concern for hybrid robotic systems.
Existing safety-critical control approaches for hybrid systems are either computationally inefficient, detrimental to system performance, or limited to small-scale systems.
We propose a learningenabled approach to construct local Control Barrier Functions (CBFs) to guarantee the safety of a wide class of nonlinear hybrid dynamical systems.
- Score: 11.57209279619218
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Hybrid dynamical systems are ubiquitous as practical robotic applications
often involve both continuous states and discrete switchings. Safety is a
primary concern for hybrid robotic systems. Existing safety-critical control
approaches for hybrid systems are either computationally inefficient,
detrimental to system performance, or limited to small-scale systems. To amend
these drawbacks, in this paper, we propose a learningenabled approach to
construct local Control Barrier Functions (CBFs) to guarantee the safety of a
wide class of nonlinear hybrid dynamical systems. The end result is a safe
neural CBFbased switching controller. Our approach is computationally
efficient, minimally invasive to any reference controller, and applicable to
large-scale systems. We empirically evaluate our framework and demonstrate its
efficacy and flexibility through two robotic examples including a
high-dimensional autonomous racing case, against other CBF-based approaches and
model predictive control.
Related papers
- Pareto Control Barrier Function for Inner Safe Set Maximization Under Input Constraints [50.920465513162334]
We introduce the PCBF algorithm to maximize the inner safe set of dynamical systems under input constraints.
We validate its effectiveness through comparison with Hamilton-Jacobi reachability for an inverted pendulum and through simulations on a 12-dimensional quadrotor system.
Results show that the PCBF consistently outperforms existing methods, yielding larger safe sets and ensuring safety under input constraints.
arXiv Detail & Related papers (2024-10-05T18:45:19Z) - Fault Tolerant Neural Control Barrier Functions for Robotic Systems
under Sensor Faults and Attacks [6.314000948709254]
We study safety-critical control synthesis for robotic systems under sensor faults and attacks.
Our main contribution is the development and synthesis of a new class of CBFs that we term fault tolerant neural control barrier function (FT-NCBF)
arXiv Detail & Related papers (2024-02-28T19:44:19Z) - Safe Neural Control for Non-Affine Control Systems with Differentiable
Control Barrier Functions [58.19198103790931]
This paper addresses the problem of safety-critical control for non-affine control systems.
It has been shown that optimizing quadratic costs subject to state and control constraints can be sub-optimally reduced to a sequence of quadratic programs (QPs) by using Control Barrier Functions (CBFs)
We incorporate higher-order CBFs into neural ordinary differential equation-based learning models as differentiable CBFs to guarantee safety for non-affine control systems.
arXiv Detail & Related papers (2023-09-06T05:35:48Z) - In-Distribution Barrier Functions: Self-Supervised Policy Filters that
Avoid Out-of-Distribution States [84.24300005271185]
We propose a control filter that wraps any reference policy and effectively encourages the system to stay in-distribution with respect to offline-collected safe demonstrations.
Our method is effective for two different visuomotor control tasks in simulation environments, including both top-down and egocentric view settings.
arXiv Detail & Related papers (2023-01-27T22:28:19Z) - Recursively Feasible Probabilistic Safe Online Learning with Control Barrier Functions [60.26921219698514]
We introduce a model-uncertainty-aware reformulation of CBF-based safety-critical controllers.
We then present the pointwise feasibility conditions of the resulting safety controller.
We use these conditions to devise an event-triggered online data collection strategy.
arXiv Detail & Related papers (2022-08-23T05:02:09Z) - Bridging Model-based Safety and Model-free Reinforcement Learning
through System Identification of Low Dimensional Linear Models [16.511440197186918]
We propose a new method to combine model-based safety with model-free reinforcement learning.
We show that a low-dimensional dynamical model is sufficient to capture the dynamics of the closed-loop system.
We illustrate that the found linear model is able to provide guarantees by safety-critical optimal control framework.
arXiv Detail & Related papers (2022-05-11T22:03:18Z) - A Barrier Certificate-based Simplex Architecture for Systems with Approximate and Hybrid Dynamics [3.254402439041394]
Bb-Simplex is a new design for runtime assurance of continuous dynamical systems.
Barrier certificates are used to prove that the baseline controller ensures safety.
Bb-Simplex can automatically derive switching conditions for complex continuous-time and hybrid systems.
arXiv Detail & Related papers (2022-02-20T02:08:10Z) - Learning Robust Hybrid Control Barrier Functions for Uncertain Systems [68.30783663518821]
We propose robust hybrid control barrier functions as a means to synthesize control laws that ensure robust safety.
Based on this notion, we formulate an optimization problem for learning robust hybrid control barrier functions from data.
Our techniques allow us to safely expand the region of attraction of a compass gait walker that is subject to model uncertainty.
arXiv Detail & Related papers (2021-01-16T17:53:35Z) - Learning Hybrid Control Barrier Functions from Data [66.37785052099423]
Motivated by the lack of systematic tools to obtain safe control laws for hybrid systems, we propose an optimization-based framework for learning certifiably safe control laws from data.
In particular, we assume a setting in which the system dynamics are known and in which data exhibiting safe system behavior is available.
arXiv Detail & Related papers (2020-11-08T23:55:02Z) - Neural Lyapunov Redesign [36.2939747271983]
Learning controllers must guarantee some notion of safety to ensure that it does not harm either the agent or the environment.
Lyapunov functions are effective tools to assess stability in nonlinear dynamical systems.
We propose a two-player collaborative algorithm that alternates between estimating a Lyapunov function and deriving a controller that gradually enlarges the stability region.
arXiv Detail & Related papers (2020-06-06T19:22:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.