Learning Stabilizing Controllers for Unstable Linear Quadratic
Regulators from a Single Trajectory
- URL: http://arxiv.org/abs/2006.11022v2
- Date: Mon, 23 Nov 2020 09:40:25 GMT
- Title: Learning Stabilizing Controllers for Unstable Linear Quadratic
Regulators from a Single Trajectory
- Authors: Lenart Treven, Sebastian Curi, Mojmir Mutny, Andreas Krause
- Abstract summary: We study linear controllers under quadratic costs model also known as linear quadratic regulators (LQR)
We present two different semi-definite programs (SDP) which results in a controller that stabilizes all systems within an ellipsoid uncertainty set.
We propose an efficient data dependent algorithm -- textsceXploration -- that with high probability quickly identifies a stabilizing controller.
- Score: 85.29718245299341
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The principal task to control dynamical systems is to ensure their stability.
When the system is unknown, robust approaches are promising since they aim to
stabilize a large set of plausible systems simultaneously. We study linear
controllers under quadratic costs model also known as linear quadratic
regulators (LQR). We present two different semi-definite programs (SDP) which
results in a controller that stabilizes all systems within an ellipsoid
uncertainty set. We further show that the feasibility conditions of the
proposed SDPs are \emph{equivalent}. Using the derived robust controller
syntheses, we propose an efficient data dependent algorithm --
\textsc{eXploration} -- that with high probability quickly identifies a
stabilizing controller. Our approach can be used to initialize existing
algorithms that require a stabilizing controller as an input while adding
constant to the regret. We further propose different heuristics which
empirically reduce the number of steps taken by \textsc{eXploration} and reduce
the suffered cost while searching for a stabilizing controller.
Related papers
- Stability Bounds for Learning-Based Adaptive Control of Discrete-Time
Multi-Dimensional Stochastic Linear Systems with Input Constraints [3.8004168340068336]
We consider the problem of adaptive stabilization for discrete-time, multi-dimensional systems with bounded control input constraints and unbounded disturbances.
We propose a certainty-equivalent control scheme which combines online parameter estimation with saturated linear control.
arXiv Detail & Related papers (2023-04-02T16:38:13Z) - Neural System Level Synthesis: Learning over All Stabilizing Policies
for Nonlinear Systems [0.0]
We propose a Neural SLS (Neur-SLS) approach guaranteeing closed-loop stability during and after parameter optimization.
We exploit recent Deep Neural Network (DNN) models based on Recurrent Equilibrium Networks (RENs) to learn over a rich class of nonlinear stable operators.
arXiv Detail & Related papers (2022-03-22T15:22:31Z) - Stabilizing Dynamical Systems via Policy Gradient Methods [32.88312419270879]
We provide a model-free algorithm for stabilizing fully observed dynamical systems.
We prove that this method efficiently recovers a stabilizing controller for linear systems.
We empirically evaluate the effectiveness of our approach on common control benchmarks.
arXiv Detail & Related papers (2021-10-13T00:58:57Z) - Finite-time System Identification and Adaptive Control in Autoregressive
Exogenous Systems [79.67879934935661]
We study the problem of system identification and adaptive control of unknown ARX systems.
We provide finite-time learning guarantees for the ARX systems under both open-loop and closed-loop data collection.
arXiv Detail & Related papers (2021-08-26T18:00:00Z) - Stable Online Control of Linear Time-Varying Systems [49.41696101740271]
COCO-LQ is an efficient online control algorithm that guarantees input-to-state stability for a large class of LTV systems.
We empirically demonstrate the performance of COCO-LQ in both synthetic experiments and a power system frequency control example.
arXiv Detail & Related papers (2021-04-29T06:18:49Z) - Gaussian Process-based Min-norm Stabilizing Controller for
Control-Affine Systems with Uncertain Input Effects and Dynamics [90.81186513537777]
We propose a novel compound kernel that captures the control-affine nature of the problem.
We show that this resulting optimization problem is convex, and we call it Gaussian Process-based Control Lyapunov Function Second-Order Cone Program (GP-CLF-SOCP)
arXiv Detail & Related papers (2020-11-14T01:27:32Z) - Robust Model-Free Learning and Control without Prior Knowledge [1.14219428942199]
We present a model-free control algorithm that robustly learn and stabilize an unknown discrete-time linear system.
The controller does not require any prior knowledge of the system dynamics, disturbances, or noise.
We will conclude with simulation results that show that despite the generality and simplicity, the controller demonstrates good closed-loop performance.
arXiv Detail & Related papers (2020-10-01T05:43:33Z) - Adaptive Control and Regret Minimization in Linear Quadratic Gaussian
(LQG) Setting [91.43582419264763]
We propose LqgOpt, a novel reinforcement learning algorithm based on the principle of optimism in the face of uncertainty.
LqgOpt efficiently explores the system dynamics, estimates the model parameters up to their confidence interval, and deploys the controller of the most optimistic model.
arXiv Detail & Related papers (2020-03-12T19:56:38Z) - Improper Learning for Non-Stochastic Control [78.65807250350755]
We consider the problem of controlling a possibly unknown linear dynamical system with adversarial perturbations, adversarially chosen convex loss functions, and partially observed states.
Applying online descent to this parametrization yields a new controller which attains sublinear regret vs. a large class of closed-loop policies.
Our bounds are the first in the non-stochastic control setting that compete with emphall stabilizing linear dynamical controllers.
arXiv Detail & Related papers (2020-01-25T02:12:48Z) - Optimistic robust linear quadratic dual control [4.94950858749529]
We present a dual control strategy that attempts to combine the performance of certainty equivalence, with the practical utility of robustness.
The formulation preserves structure in the representation of parametric uncertainty, which allows the controller to target reduction of uncertainty in the parameters that matter most' for the control task.
arXiv Detail & Related papers (2019-12-31T02:02:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.