Safe Learning of Uncertain Environments for Nonlinear Control-Affine
Systems
- URL: http://arxiv.org/abs/2103.01413v1
- Date: Tue, 2 Mar 2021 01:58:02 GMT
- Title: Safe Learning of Uncertain Environments for Nonlinear Control-Affine
Systems
- Authors: Farhad Farokhi, Alex Leong, Iman Shames, Mohammad Zamani
- Abstract summary: We consider the problem of safe learning in nonlinear control-affine systems subject to unknown additive uncertainty.
We model uncertainty as a Gaussian signal and use state measurements to learn its mean and covariance bounds.
We show that with an arbitrarily large probability we can guarantee that the state will remain in the safe set, while learning and control are carried out simultaneously.
- Score: 10.918870296899245
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In many learning based control methodologies, learning the unknown dynamic
model precedes the control phase, while the aim is to control the system such
that it remains in some safe region of the state space. In this work our aim is
to guarantee safety while learning and control proceed simultaneously.
Specifically, we consider the problem of safe learning in nonlinear
control-affine systems subject to unknown additive uncertainty. We model
uncertainty as a Gaussian signal and use state measurements to learn its mean
and covariance. We provide rigorous time-varying bounds on the mean and
covariance of the uncertainty and employ them to modify the control input via
an optimisation program with safety constraints encoded as a barrier function
on the state space. We show that with an arbitrarily large probability we can
guarantee that the state will remain in the safe set, while learning and
control are carried out simultaneously, provided that a feasible solution
exists for the optimisation problem. We provide a secondary formulation of this
optimisation that is computationally more efficient. This is based on
tightening the safety constraints to counter the uncertainty about the learned
mean and covariance. The magnitude of the tightening can be decreased as our
confidence in the learned mean and covariance increases (i.e., as we gather
more measurements about the environment). Extensions of the method are provided
for Gaussian uncertainties with piecewise constant mean and covariance to
accommodate more general environments.
Related papers
- Distributionally Safe Reinforcement Learning under Model Uncertainty: A
Single-Level Approach by Differentiable Convex Programming [4.825619788907192]
We present a tractable distributionally safe reinforcement learning framework to enforce safety under a distributional shift measured by a Wasserstein metric.
To improve the tractability, we first use duality theory to transform the lower-level optimization from infinite-dimensional probability space to a finite-dimensional parametric space.
By differentiable convex programming, the bi-level safe learning problem is further reduced to a single-level one with two sequential computationally efficient modules.
arXiv Detail & Related papers (2023-10-03T22:05:05Z) - Adaptive Robust Model Predictive Control via Uncertainty Cancellation [25.736296938185074]
We propose a learning-based robust predictive control algorithm that compensates for significant uncertainty in the dynamics.
We optimize over a class of nonlinear feedback policies inspired by certainty equivalent "estimate-and-cancel" control laws.
arXiv Detail & Related papers (2022-12-02T18:54:23Z) - Enforcing Hard Constraints with Soft Barriers: Safe Reinforcement
Learning in Unknown Stochastic Environments [84.3830478851369]
We propose a safe reinforcement learning approach that can jointly learn the environment and optimize the control policy.
Our approach can effectively enforce hard safety constraints and significantly outperform CMDP-based baseline methods in system safe rate measured via simulations.
arXiv Detail & Related papers (2022-09-29T20:49:25Z) - Recursively Feasible Probabilistic Safe Online Learning with Control Barrier Functions [60.26921219698514]
We introduce a model-uncertainty-aware reformulation of CBF-based safety-critical controllers.
We then present the pointwise feasibility conditions of the resulting safety controller.
We use these conditions to devise an event-triggered online data collection strategy.
arXiv Detail & Related papers (2022-08-23T05:02:09Z) - Learning Safety Filters for Unknown Discrete-Time Linear Systems [11.533793543850384]
Safety is characterized using polytopic constraints on the states and control inputs.
The empirically learned model and process noise covariance with their confidence bounds are used to construct a robust optimization problem for minimally modifying nominal control actions to ensure safety with high probability.
arXiv Detail & Related papers (2021-11-01T00:14:08Z) - Pointwise Feasibility of Gaussian Process-based Safety-Critical Control
under Model Uncertainty [77.18483084440182]
Control Barrier Functions (CBFs) and Control Lyapunov Functions (CLFs) are popular tools for enforcing safety and stability of a controlled system, respectively.
We present a Gaussian Process (GP)-based approach to tackle the problem of model uncertainty in safety-critical controllers that use CBFs and CLFs.
arXiv Detail & Related papers (2021-06-13T23:08:49Z) - Adaptive Robust Model Predictive Control with Matched and Unmatched
Uncertainty [28.10549712956161]
We propose a learning-based robust predictive control algorithm that can handle large uncertainty in the dynamics for a class of discrete-time systems.
Motivated by an inability of existing learning-based predictive control algorithms to achieve safety guarantees in the presence of uncertainties of large magnitude, we achieve significant performance improvements.
arXiv Detail & Related papers (2021-04-16T17:47:02Z) - Closing the Closed-Loop Distribution Shift in Safe Imitation Learning [80.05727171757454]
We treat safe optimization-based control strategies as experts in an imitation learning problem.
We train a learned policy that can be cheaply evaluated at run-time and that provably satisfies the same safety guarantees as the expert.
arXiv Detail & Related papers (2021-02-18T05:11:41Z) - Improper Learning with Gradient-based Policy Optimization [62.50997487685586]
We consider an improper reinforcement learning setting where the learner is given M base controllers for an unknown Markov Decision Process.
We propose a gradient-based approach that operates over a class of improper mixtures of the controllers.
arXiv Detail & Related papers (2021-02-16T14:53:55Z) - Chance-Constrained Trajectory Optimization for Safe Exploration and
Learning of Nonlinear Systems [81.7983463275447]
Learning-based control algorithms require data collection with abundant supervision for training.
We present a new approach for optimal motion planning with safe exploration that integrates chance-constrained optimal control with dynamics learning and feedback control.
arXiv Detail & Related papers (2020-05-09T05:57:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.