Robust Control for Dynamical Systems With Non-Gaussian Noise via Formal
Abstractions
- URL: http://arxiv.org/abs/2301.01526v1
- Date: Wed, 4 Jan 2023 10:40:30 GMT
- Title: Robust Control for Dynamical Systems With Non-Gaussian Noise via Formal
Abstractions
- Authors: Thom Badings and Licio Romao and Alessandro Abate and David Parker and
Hasan A. Poonawala and Marielle Stoelinga and Nils Jansen
- Abstract summary: We present a novel controller synthesis method that does not rely on any explicit representation of the noise distributions.
First, we abstract the continuous control system into a finite-state model that captures noise by probabilistic transitions between discrete states.
We use state-of-the-art verification techniques to provide guarantees on the interval Markov decision process and compute a controller for which these guarantees carry over to the original control system.
- Score: 59.605246463200736
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Controllers for dynamical systems that operate in safety-critical settings
must account for stochastic disturbances. Such disturbances are often modeled
as process noise in a dynamical system, and common assumptions are that the
underlying distributions are known and/or Gaussian. In practice, however, these
assumptions may be unrealistic and can lead to poor approximations of the true
noise distribution. We present a novel controller synthesis method that does
not rely on any explicit representation of the noise distributions. In
particular, we address the problem of computing a controller that provides
probabilistic guarantees on safely reaching a target, while also avoiding
unsafe regions of the state space. First, we abstract the continuous control
system into a finite-state model that captures noise by probabilistic
transitions between discrete states. As a key contribution, we adapt tools from
the scenario approach to compute probably approximately correct (PAC) bounds on
these transition probabilities, based on a finite number of samples of the
noise. We capture these bounds in the transition probability intervals of a
so-called interval Markov decision process (iMDP). This iMDP is, with a
user-specified confidence probability, robust against uncertainty in the
transition probabilities, and the tightness of the probability intervals can be
controlled through the number of samples. We use state-of-the-art verification
techniques to provide guarantees on the iMDP and compute a controller for which
these guarantees carry over to the original control system. In addition, we
develop a tailored computational scheme that reduces the complexity of the
synthesis of these guarantees on the iMDP. Benchmarks on realistic control
systems show the practical applicability of our method, even when the iMDP has
hundreds of millions of transitions.
Related papers
- Formal Controller Synthesis for Markov Jump Linear Systems with
Uncertain Dynamics [64.72260320446158]
We propose a method for synthesising controllers for Markov jump linear systems.
Our method is based on a finite-state abstraction that captures both the discrete (mode-jumping) and continuous (stochastic linear) behaviour of the MJLS.
We apply our method to multiple realistic benchmark problems, in particular, a temperature control and an aerial vehicle delivery problem.
arXiv Detail & Related papers (2022-12-01T17:36:30Z) - Probabilities Are Not Enough: Formal Controller Synthesis for Stochastic
Dynamical Models with Epistemic Uncertainty [68.00748155945047]
Capturing uncertainty in models of complex dynamical systems is crucial to designing safe controllers.
Several approaches use formal abstractions to synthesize policies that satisfy temporal specifications related to safety and reachability.
Our contribution is a novel abstraction-based controller method for continuous-state models with noise, uncertain parameters, and external disturbances.
arXiv Detail & Related papers (2022-10-12T07:57:03Z) - Adaptive Stochastic MPC under Unknown Noise Distribution [19.03553854357296]
We address the MPC problem for linear systems, subject to chance state constraints and hard input constraints, under unknown noise distribution.
We design a distributionally robust and robustly stable benchmark SMPC algorithm for the ideal setting of known noise statistics.
We employ this benchmark controller to derive a novel adaptive SMPC scheme that learns the necessary noise statistics online.
arXiv Detail & Related papers (2022-04-03T16:35:18Z) - Sampling-Based Robust Control of Autonomous Systems with Non-Gaussian
Noise [59.47042225257565]
We present a novel planning method that does not rely on any explicit representation of the noise distributions.
First, we abstract the continuous system into a discrete-state model that captures noise by probabilistic transitions between states.
We capture these bounds in the transition probability intervals of a so-called interval Markov decision process (iMDP)
arXiv Detail & Related papers (2021-10-25T06:18:55Z) - Pointwise Feasibility of Gaussian Process-based Safety-Critical Control
under Model Uncertainty [77.18483084440182]
Control Barrier Functions (CBFs) and Control Lyapunov Functions (CLFs) are popular tools for enforcing safety and stability of a controlled system, respectively.
We present a Gaussian Process (GP)-based approach to tackle the problem of model uncertainty in safety-critical controllers that use CBFs and CLFs.
arXiv Detail & Related papers (2021-06-13T23:08:49Z) - Probabilistic robust linear quadratic regulators with Gaussian processes [73.0364959221845]
Probabilistic models such as Gaussian processes (GPs) are powerful tools to learn unknown dynamical systems from data for subsequent use in control design.
We present a novel controller synthesis for linearized GP dynamics that yields robust controllers with respect to a probabilistic stability margin.
arXiv Detail & Related papers (2021-05-17T08:36:18Z) - Correct-by-construction reach-avoid control of partially observable
linear stochastic systems [7.912008109232803]
We formalize a robust feedback controller for reach-avoid control of discrete-time, linear time-invariant systems.
The problem is to compute a controller that satisfies the required provestate abstraction problem.
arXiv Detail & Related papers (2021-03-03T13:46:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.