Safely Learning Dynamical Systems from Short Trajectories
- URL: http://arxiv.org/abs/2011.12257v1
- Date: Tue, 24 Nov 2020 18:06:10 GMT
- Title: Safely Learning Dynamical Systems from Short Trajectories
- Authors: Amir Ali Ahmadi, Abraar Chaudhry, Vikas Sindhwani, Stephen Tu
- Abstract summary: A fundamental challenge in learning to control an unknown dynamical system is to reduce model uncertainty by making measurements while maintaining safety.
We formulate a mathematical definition of what it means to safely learn a dynamical system by sequentially deciding where to initialize the next trajectory.
We present a linear programming-based algorithm that either safely recovers the true dynamics from trajectories of length one, or certifies that safe learning is impossible.
- Score: 12.184674552836414
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: A fundamental challenge in learning to control an unknown dynamical system is
to reduce model uncertainty by making measurements while maintaining safety. In
this work, we formulate a mathematical definition of what it means to safely
learn a dynamical system by sequentially deciding where to initialize the next
trajectory. In our framework, the state of the system is required to stay
within a given safety region under the (possibly repeated) action of all
dynamical systems that are consistent with the information gathered so far. For
our first two results, we consider the setting of safely learning linear
dynamics. We present a linear programming-based algorithm that either safely
recovers the true dynamics from trajectories of length one, or certifies that
safe learning is impossible. We also give an efficient semidefinite
representation of the set of initial conditions whose resulting trajectories of
length two are guaranteed to stay in the safety region. For our final result,
we study the problem of safely learning a nonlinear dynamical system. We give a
second-order cone programming based representation of the set of initial
conditions that are guaranteed to remain in the safety region after one
application of the system dynamics.
Related papers
- Neural Contractive Dynamical Systems [13.046426079291376]
Stability guarantees are crucial when ensuring a fully autonomous robot does not take undesirable or potentially harmful actions.
We propose a novel methodology to learn neural contractive dynamical systems, where our neural architecture ensures contraction.
We show that our approach encodes the desired dynamics more accurately than the current state-of-the-art, which provides less strong stability guarantees.
arXiv Detail & Related papers (2024-01-17T17:18:21Z) - Safely Learning Dynamical Systems [13.598953823625472]
A fundamental challenge in learning an unknown dynamical system is to reduce model uncertainty by making measurements while maintaining safety.
We formulate a mathematical definition of what it means to safely learn a dynamical system by sequentially deciding where to initialize trajectories.
We show how one can safely collect trajectories and fit a model of the nonlinear dynamics that is consistent with the initial uncertainty set.
arXiv Detail & Related papers (2023-05-20T21:35:29Z) - Statistical Safety and Robustness Guarantees for Feedback Motion
Planning of Unknown Underactuated Stochastic Systems [1.0323063834827415]
We propose a sampling-based planner that uses the mean dynamics model and simultaneously bounds the closed-loop tracking error via a learned disturbance bound.
We validate that our guarantees translate to empirical safety in simulation on a 10D quadrotor, and in the real world on a physical CrazyFlie quadrotor and Clearpath Jackal robot.
arXiv Detail & Related papers (2022-12-13T19:38:39Z) - Recursively Feasible Probabilistic Safe Online Learning with Control Barrier Functions [60.26921219698514]
We introduce a model-uncertainty-aware reformulation of CBF-based safety-critical controllers.
We then present the pointwise feasibility conditions of the resulting safety controller.
We use these conditions to devise an event-triggered online data collection strategy.
arXiv Detail & Related papers (2022-08-23T05:02:09Z) - Learning Barrier Certificates: Towards Safe Reinforcement Learning with
Zero Training-time Violations [64.39401322671803]
This paper explores the possibility of safe RL algorithms with zero training-time safety violations.
We propose an algorithm, Co-trained Barrier Certificate for Safe RL (CRABS), which iteratively learns barrier certificates, dynamics models, and policies.
arXiv Detail & Related papers (2021-08-04T04:59:05Z) - Safe Active Dynamics Learning and Control: A Sequential
Exploration-Exploitation Framework [30.58186749790728]
We propose a theoretically-justified approach to maintaining safety in the presence of dynamics uncertainty.
Our framework guarantees the high-probability satisfaction of all constraints at all times jointly.
This theoretical analysis also motivates two regularizers of last-layer meta-learning models that improve online adaptation capabilities.
arXiv Detail & Related papers (2020-08-26T17:39:58Z) - Reinforcement Learning with Fast Stabilization in Linear Dynamical
Systems [91.43582419264763]
We study model-based reinforcement learning (RL) in unknown stabilizable linear dynamical systems.
We propose an algorithm that certifies fast stabilization of the underlying system by effectively exploring the environment.
We show that the proposed algorithm attains $tildemathcalO(sqrtT)$ regret after $T$ time steps of agent-environment interaction.
arXiv Detail & Related papers (2020-07-23T23:06:40Z) - Active Learning for Nonlinear System Identification with Guarantees [102.43355665393067]
We study a class of nonlinear dynamical systems whose state transitions depend linearly on a known feature embedding of state-action pairs.
We propose an active learning approach that achieves this by repeating three steps: trajectory planning, trajectory tracking, and re-estimation of the system from all available data.
We show that our method estimates nonlinear dynamical systems at a parametric rate, similar to the statistical rate of standard linear regression.
arXiv Detail & Related papers (2020-06-18T04:54:11Z) - Neural Certificates for Safe Control Policies [108.4560749465701]
This paper develops an approach to learn a policy of a dynamical system that is guaranteed to be both provably safe and goal-reaching.
We show the effectiveness of the method to learn both safe and goal-reaching policies on various systems, including pendulums, cart-poles, and UAVs.
arXiv Detail & Related papers (2020-06-15T15:14:18Z) - Chance-Constrained Trajectory Optimization for Safe Exploration and
Learning of Nonlinear Systems [81.7983463275447]
Learning-based control algorithms require data collection with abundant supervision for training.
We present a new approach for optimal motion planning with safe exploration that integrates chance-constrained optimal control with dynamics learning and feedback control.
arXiv Detail & Related papers (2020-05-09T05:57:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.