Planning with Learned Dynamics: Probabilistic Guarantees on Safety and
Reachability via Lipschitz Constants
- URL: http://arxiv.org/abs/2010.08993v4
- Date: Tue, 19 Oct 2021 18:46:13 GMT
- Title: Planning with Learned Dynamics: Probabilistic Guarantees on Safety and
Reachability via Lipschitz Constants
- Authors: Craig Knuth, Glen Chou, Necmiye Ozay, Dmitry Berenson
- Abstract summary: We present a method for feedback motion planning of systems with unknown dynamics.
We provide guarantees on safety, reachability, and goal stability.
We demonstrate our approach by planning using learned models of a 6D quadrotor and a 7DOF Kuka arm.
- Score: 7.216586291939535
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present a method for feedback motion planning of systems with unknown
dynamics which provides probabilistic guarantees on safety, reachability, and
goal stability. To find a domain in which a learned control-affine
approximation of the true dynamics can be trusted, we estimate the Lipschitz
constant of the difference between the true and learned dynamics, and ensure
the estimate is valid with a given probability. Provided the system has at
least as many controls as states, we also derive existence conditions for a
one-step feedback law which can keep the real system within a small bound of a
nominal trajectory planned with the learned dynamics. Our method imposes the
feedback law existence as a constraint in a sampling-based planner, which
returns a feedback policy around a nominal plan ensuring that, if the Lipschitz
constant estimate is valid, the true system is safe during plan execution,
reaches the goal, and is ultimately invariant in a small set about the goal. We
demonstrate our approach by planning using learned models of a 6D quadrotor and
a 7DOF Kuka arm. We show that a baseline which plans using the same learned
dynamics without considering the error bound or the existence of the feedback
law can fail to stabilize around the plan and become unsafe.
Related papers
- Probabilistic Reach-Avoid for Bayesian Neural Networks [71.67052234622781]
We show that an optimal synthesis algorithm can provide more than a four-fold increase in the number of certifiable states.
The algorithm is able to provide more than a three-fold increase in the average guaranteed reach-avoid probability.
arXiv Detail & Related papers (2023-10-03T10:52:21Z) - Statistical Safety and Robustness Guarantees for Feedback Motion
Planning of Unknown Underactuated Stochastic Systems [1.0323063834827415]
We propose a sampling-based planner that uses the mean dynamics model and simultaneously bounds the closed-loop tracking error via a learned disturbance bound.
We validate that our guarantees translate to empirical safety in simulation on a 10D quadrotor, and in the real world on a physical CrazyFlie quadrotor and Clearpath Jackal robot.
arXiv Detail & Related papers (2022-12-13T19:38:39Z) - Recursively Feasible Probabilistic Safe Online Learning with Control Barrier Functions [60.26921219698514]
We introduce a model-uncertainty-aware reformulation of CBF-based safety-critical controllers.
We then present the pointwise feasibility conditions of the resulting safety controller.
We use these conditions to devise an event-triggered online data collection strategy.
arXiv Detail & Related papers (2022-08-23T05:02:09Z) - Safe Reinforcement Learning via Confidence-Based Filters [78.39359694273575]
We develop a control-theoretic approach for certifying state safety constraints for nominal policies learned via standard reinforcement learning techniques.
We provide formal safety guarantees, and empirically demonstrate the effectiveness of our approach.
arXiv Detail & Related papers (2022-07-04T11:43:23Z) - ProBF: Learning Probabilistic Safety Certificates with Barrier Functions [31.203344483485843]
The control barrier function is a useful tool to guarantee safety if we have access to the ground-truth system dynamics.
In practice, we have inaccurate knowledge of the system dynamics, which can lead to unsafe behaviors.
We show the efficacy of this method through experiments on Segway and Quadrotor simulations.
arXiv Detail & Related papers (2021-12-22T20:18:18Z) - Probabilistic robust linear quadratic regulators with Gaussian processes [73.0364959221845]
Probabilistic models such as Gaussian processes (GPs) are powerful tools to learn unknown dynamical systems from data for subsequent use in control design.
We present a novel controller synthesis for linearized GP dynamics that yields robust controllers with respect to a probabilistic stability margin.
arXiv Detail & Related papers (2021-05-17T08:36:18Z) - Model Error Propagation via Learned Contraction Metrics for Safe
Feedback Motion Planning of Unknown Systems [4.702729080310267]
We present a method for contraction-based feedback motion planning of locally incrementally stabilizable systems with unknown dynamics.
Given a dynamics dataset, our method learns a deep control-affine approximation of the dynamics.
We show results on a 4D car, a 6D quadrotor, and a 22D deformable object manipulation task, showing our method plans safely with learned models of high-dimensional underactuated systems.
arXiv Detail & Related papers (2021-04-18T03:34:00Z) - Safe Learning of Uncertain Environments for Nonlinear Control-Affine
Systems [10.918870296899245]
We consider the problem of safe learning in nonlinear control-affine systems subject to unknown additive uncertainty.
We model uncertainty as a Gaussian signal and use state measurements to learn its mean and covariance bounds.
We show that with an arbitrarily large probability we can guarantee that the state will remain in the safe set, while learning and control are carried out simultaneously.
arXiv Detail & Related papers (2021-03-02T01:58:02Z) - Uncertainty-Aware Constraint Learning for Adaptive Safe Motion Planning
from Demonstrations [6.950510860295866]
We present a method for learning to satisfy uncertain constraints from demonstrations.
Our method uses robust optimization to obtain a belief over the potentially infinite set of possible constraints consistent with the demonstrations.
We derive guarantees on the accuracy of our constraint belief and probabilistic guarantees on plan safety.
arXiv Detail & Related papers (2020-11-09T01:59:14Z) - Chance-Constrained Trajectory Optimization for Safe Exploration and
Learning of Nonlinear Systems [81.7983463275447]
Learning-based control algorithms require data collection with abundant supervision for training.
We present a new approach for optimal motion planning with safe exploration that integrates chance-constrained optimal control with dynamics learning and feedback control.
arXiv Detail & Related papers (2020-05-09T05:57:43Z) - Learning Control Barrier Functions from Expert Demonstrations [69.23675822701357]
We propose a learning based approach to safe controller synthesis based on control barrier functions (CBFs)
We analyze an optimization-based approach to learning a CBF that enjoys provable safety guarantees under suitable Lipschitz assumptions on the underlying dynamical system.
To the best of our knowledge, these are the first results that learn provably safe control barrier functions from data.
arXiv Detail & Related papers (2020-04-07T12:29:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.