Safety-Critical Online Control with Adversarial Disturbances
- URL: http://arxiv.org/abs/2009.09511v1
- Date: Sun, 20 Sep 2020 19:59:15 GMT
- Title: Safety-Critical Online Control with Adversarial Disturbances
- Authors: Bhaskar Ramasubramanian, Baicen Xiao, Linda Bushnell, Radha Poovendran
- Abstract summary: We seek to synthesize state-feedback controllers to minimize a cost incurred due to the disturbance.
We consider an online setting where costs at each time are revealed only after the controller at that time is chosen.
We show that the regret function, which is defined as the difference between these costs, varies logarithmically with the time horizon.
- Score: 8.633140051496408
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper studies the control of safety-critical dynamical systems in the
presence of adversarial disturbances. We seek to synthesize state-feedback
controllers to minimize a cost incurred due to the disturbance, while
respecting a safety constraint. The safety constraint is given by a bound on an
H-inf norm, while the cost is specified as an upper bound on the H-2 norm of
the system. We consider an online setting where costs at each time are revealed
only after the controller at that time is chosen. We propose an iterative
approach to the synthesis of the controller by solving a modified discrete-time
Riccati equation. Solutions of this equation enforce the safety constraint. We
compare the cost of this controller with that of the optimal controller when
one has complete knowledge of disturbances and costs in hindsight. We show that
the regret function, which is defined as the difference between these costs,
varies logarithmically with the time horizon. We validate our approach on a
process control setup that is subject to two kinds of adversarial attacks.
Related papers
- Safe Neural Control for Non-Affine Control Systems with Differentiable
Control Barrier Functions [58.19198103790931]
This paper addresses the problem of safety-critical control for non-affine control systems.
It has been shown that optimizing quadratic costs subject to state and control constraints can be sub-optimally reduced to a sequence of quadratic programs (QPs) by using Control Barrier Functions (CBFs)
We incorporate higher-order CBFs into neural ordinary differential equation-based learning models as differentiable CBFs to guarantee safety for non-affine control systems.
arXiv Detail & Related papers (2023-09-06T05:35:48Z) - Regret-optimal Estimation and Control [52.28457815067461]
We show that the regret-optimal estimator and regret-optimal controller can be derived in state-space form.
We propose regret-optimal analogs of Model-Predictive Control (MPC) and the Extended KalmanFilter (EKF) for systems with nonlinear dynamics.
arXiv Detail & Related papers (2021-06-22T23:14:21Z) - Stable Online Control of Linear Time-Varying Systems [49.41696101740271]
COCO-LQ is an efficient online control algorithm that guarantees input-to-state stability for a large class of LTV systems.
We empirically demonstrate the performance of COCO-LQ in both synthetic experiments and a power system frequency control example.
arXiv Detail & Related papers (2021-04-29T06:18:49Z) - Non-stationary Online Learning with Memory and Non-stochastic Control [71.14503310914799]
We study the problem of Online Convex Optimization (OCO) with memory, which allows loss functions to depend on past decisions.
In this paper, we introduce dynamic policy regret as the performance measure to design algorithms robust to non-stationary environments.
We propose a novel algorithm for OCO with memory that provably enjoys an optimal dynamic policy regret in terms of time horizon, non-stationarity measure, and memory length.
arXiv Detail & Related papers (2021-02-07T09:45:15Z) - Meta-Learning Guarantees for Online Receding Horizon Learning Control [0.0]
We provide provable regret guarantees for an online meta-learning receding horizon control algorithm in an iterative control setting.
We show that the worst regret for learning within an iteration improves with experience of more iterations.
arXiv Detail & Related papers (2020-10-21T21:57:04Z) - A Meta-Learning Control Algorithm with Provable Finite-Time Guarantees [0.0]
We provide provable regret guarantees for an online meta-learning control algorithm in an iterative control setting.
We show that the worst regret for the learning within an iteration continuously improves with experience of more iterations.
arXiv Detail & Related papers (2020-08-30T20:30:40Z) - Learning Stabilizing Controllers for Unstable Linear Quadratic
Regulators from a Single Trajectory [85.29718245299341]
We study linear controllers under quadratic costs model also known as linear quadratic regulators (LQR)
We present two different semi-definite programs (SDP) which results in a controller that stabilizes all systems within an ellipsoid uncertainty set.
We propose an efficient data dependent algorithm -- textsceXploration -- that with high probability quickly identifies a stabilizing controller.
arXiv Detail & Related papers (2020-06-19T08:58:57Z) - Improper Learning for Non-Stochastic Control [78.65807250350755]
We consider the problem of controlling a possibly unknown linear dynamical system with adversarial perturbations, adversarially chosen convex loss functions, and partially observed states.
Applying online descent to this parametrization yields a new controller which attains sublinear regret vs. a large class of closed-loop policies.
Our bounds are the first in the non-stochastic control setting that compete with emphall stabilizing linear dynamical controllers.
arXiv Detail & Related papers (2020-01-25T02:12:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.