Reinforcement Learning for Safety-Critical Control under Model
Uncertainty, using Control Lyapunov Functions and Control Barrier Functions
- URL: http://arxiv.org/abs/2004.07584v2
- Date: Thu, 4 Jun 2020 17:07:52 GMT
- Title: Reinforcement Learning for Safety-Critical Control under Model
Uncertainty, using Control Lyapunov Functions and Control Barrier Functions
- Authors: Jason Choi, Fernando Casta\~neda, Claire J. Tomlin, Koushil Sreenath
- Abstract summary: Reinforcement learning framework learns the model uncertainty present in the CBF and CLF constraints.
RL-CBF-CLF-QP addresses the problem of model uncertainty in the safety constraints.
- Score: 96.63967125746747
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, the issue of model uncertainty in safety-critical control is
addressed with a data-driven approach. For this purpose, we utilize the
structure of an input-ouput linearization controller based on a nominal model
along with a Control Barrier Function and Control Lyapunov Function based
Quadratic Program (CBF-CLF-QP). Specifically, we propose a novel reinforcement
learning framework which learns the model uncertainty present in the CBF and
CLF constraints, as well as other control-affine dynamic constraints in the
quadratic program. The trained policy is combined with the nominal model-based
CBF-CLF-QP, resulting in the Reinforcement Learning-based CBF-CLF-QP
(RL-CBF-CLF-QP), which addresses the problem of model uncertainty in the safety
constraints. The performance of the proposed method is validated by testing it
on an underactuated nonlinear bipedal robot walking on randomly spaced stepping
stones with one step preview, obtaining stable and safe walking under model
uncertainty.
Related papers
- Pareto Control Barrier Function for Inner Safe Set Maximization Under Input Constraints [50.920465513162334]
We introduce the PCBF algorithm to maximize the inner safe set of dynamical systems under input constraints.
We validate its effectiveness through comparison with Hamilton-Jacobi reachability for an inverted pendulum and through simulations on a 12-dimensional quadrotor system.
Results show that the PCBF consistently outperforms existing methods, yielding larger safe sets and ensuring safety under input constraints.
arXiv Detail & Related papers (2024-10-05T18:45:19Z) - Safe Neural Control for Non-Affine Control Systems with Differentiable
Control Barrier Functions [58.19198103790931]
This paper addresses the problem of safety-critical control for non-affine control systems.
It has been shown that optimizing quadratic costs subject to state and control constraints can be sub-optimally reduced to a sequence of quadratic programs (QPs) by using Control Barrier Functions (CBFs)
We incorporate higher-order CBFs into neural ordinary differential equation-based learning models as differentiable CBFs to guarantee safety for non-affine control systems.
arXiv Detail & Related papers (2023-09-06T05:35:48Z) - Model-Assisted Probabilistic Safe Adaptive Control With Meta-Bayesian
Learning [33.75998206184497]
We develop a novel adaptive safe control framework that integrates meta learning, Bayesian models, and control barrier function (CBF) method.
Specifically, with the help of CBF method, we learn the inherent and external uncertainties by a unified adaptive Bayesian linear regression model.
For a new control task, we refine the meta-learned models using a few samples, and introduce pessimistic confidence bounds into CBF constraints to ensure safe control.
arXiv Detail & Related papers (2023-07-03T08:16:01Z) - Stable and Safe Reinforcement Learning via a Barrier-Lyapunov
Actor-Critic Approach [1.8924647429604111]
Barrier-Lyapunov Actor-Critic (BLAC) framework helps maintain the aforementioned safety and stability for the system.
Additional backup controller is introduced in case the RL-based controller cannot provide valid control signals.
arXiv Detail & Related papers (2023-04-08T16:48:49Z) - Differentiable Predictive Control with Safety Guarantees: A Control
Barrier Function Approach [3.617866023850784]
We develop a novel form of differentiable predictive control (DPC) with safety and robustness guarantees.
DPC is an unsupervised learning-based method for obtaining approximate solutions to explicit model predictive control (MPC) problems.
arXiv Detail & Related papers (2022-08-03T19:24:44Z) - Learning Robust Output Control Barrier Functions from Safe Expert Demonstrations [50.37808220291108]
This paper addresses learning safe output feedback control laws from partial observations of expert demonstrations.
We first propose robust output control barrier functions (ROCBFs) as a means to guarantee safety.
We then formulate an optimization problem to learn ROCBFs from expert demonstrations that exhibit safe system behavior.
arXiv Detail & Related papers (2021-11-18T23:21:00Z) - Pointwise Feasibility of Gaussian Process-based Safety-Critical Control
under Model Uncertainty [77.18483084440182]
Control Barrier Functions (CBFs) and Control Lyapunov Functions (CLFs) are popular tools for enforcing safety and stability of a controlled system, respectively.
We present a Gaussian Process (GP)-based approach to tackle the problem of model uncertainty in safety-critical controllers that use CBFs and CLFs.
arXiv Detail & Related papers (2021-06-13T23:08:49Z) - Learning Control Barrier Functions from Expert Demonstrations [69.23675822701357]
We propose a learning based approach to safe controller synthesis based on control barrier functions (CBFs)
We analyze an optimization-based approach to learning a CBF that enjoys provable safety guarantees under suitable Lipschitz assumptions on the underlying dynamical system.
To the best of our knowledge, these are the first results that learn provably safe control barrier functions from data.
arXiv Detail & Related papers (2020-04-07T12:29:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.