On Regularizability and its Application to Online Control of Unstable
LTI Systems
- URL: http://arxiv.org/abs/2006.00125v3
- Date: Wed, 19 Jan 2022 21:04:25 GMT
- Title: On Regularizability and its Application to Online Control of Unstable
LTI Systems
- Authors: Shahriar Talebi, Siavash Alemzadeh, Niyousha Rahimi and Mehran Mesbahi
- Abstract summary: We examine online regulation of (possibly unstable) partially unknown linear systems with no prior access to an initial stabilizing controller or PE input-output data.
Having access only to the input matrix, we propose the Data-Guided Regulation (DGR) procedure that regulates the underlying state while also generating informative data.
We further improve the computational performance of DGR via a rank-one update and demonstrate its utility in online regulation of the X-29 aircraft.
- Score: 0.5735035463793007
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Learning, say through direct policy updates, often requires assumptions such
as knowing a priori that the initial policy (gain) is stabilizing, or
persistently exciting (PE) input-output data, is available. In this paper, we
examine online regulation of (possibly unstable) partially unknown linear
systems with no prior access to an initial stabilizing controller nor PE
input-output data; we instead leverage the knowledge of the input matrix for
online regulation. First, we introduce and characterize the notion of
"regularizability" for linear systems that gauges the extent by which a system
can be regulated in finite-time in contrast to its asymptotic behavior
(commonly characterized by stabilizability/controllability). Next, having
access only to the input matrix, we propose the Data-Guided Regulation (DGR)
synthesis procedure that -- as its name suggests -- regulates the underlying
state while also generating informative data that can subsequently be used for
data-driven stabilization or system identification. We further improve the
computational performance of DGR via a rank-one update and demonstrate its
utility in online regulation of the X-29 aircraft.
Related papers
- Implicit Bias of Policy Gradient in Linear Quadratic Control: Extrapolation to Unseen Initial States [52.56827348431552]
gradient descent frequently exhibits an implicit bias that leads to excellent performance on unseen data.
This paper theoretically studies the implicit bias of policy gradient in terms of extrapolation to unseen initial states.
arXiv Detail & Related papers (2024-02-12T18:41:31Z) - Data-Guided Regulator for Adaptive Nonlinear Control [0.27195102129094995]
This paper addresses the problem of a data-driven feedback controller for complex nonlinear systems.
The goal is to achieve finite-time regulation of system states through direct policy updates.
arXiv Detail & Related papers (2023-11-20T23:02:39Z) - Non-asymptotic System Identification for Linear Systems with Nonlinear
Policies [17.420749574975368]
This paper considers a single-trajectory system identification problem for linear systems under general nonlinear and/or time-varying policies.
We provide a non-asymptotic error bound for least square estimation when the data trajectory is generated by any nonlinear and/or time-varying policies.
arXiv Detail & Related papers (2023-06-17T15:05:59Z) - LMI-based Data-Driven Robust Model Predictive Control [0.1473281171535445]
We propose a data-driven robust linear matrix inequality-based model predictive control scheme that considers input and state constraints.
The controller stabilizes the closed-loop system and guarantees constraint satisfaction.
arXiv Detail & Related papers (2023-03-08T18:20:06Z) - Learning-Based Adaptive Control for Stochastic Linear Systems with Input
Constraints [3.8004168340068336]
We propose a certainty-equivalence scheme for adaptive control of scalar linear systems subject to additive, i.i.d.
Assuming that the system is at-worst marginally stable, mean square boundedness of the closed-loop system states is proven.
arXiv Detail & Related papers (2022-09-15T04:49:06Z) - Stable Online Control of Linear Time-Varying Systems [49.41696101740271]
COCO-LQ is an efficient online control algorithm that guarantees input-to-state stability for a large class of LTV systems.
We empirically demonstrate the performance of COCO-LQ in both synthetic experiments and a power system frequency control example.
arXiv Detail & Related papers (2021-04-29T06:18:49Z) - Controlling nonlinear dynamical systems into arbitrary states using
machine learning [77.34726150561087]
We propose a novel and fully data driven control scheme which relies on machine learning (ML)
Exploiting recently developed ML-based prediction capabilities of complex systems, we demonstrate that nonlinear systems can be forced to stay in arbitrary dynamical target states coming from any initial state.
Having this highly flexible control scheme with little demands on the amount of required data on hand, we briefly discuss possible applications that range from engineering to medicine.
arXiv Detail & Related papers (2021-02-23T16:58:26Z) - The Impact of Data on the Stability of Learning-Based Control- Extended
Version [63.97366815968177]
We propose a Lyapunov-based measure for quantifying the impact of data on the certifiable control performance.
By modeling unknown system dynamics through Gaussian processes, we can determine the interrelation between model uncertainty and satisfaction of stability conditions.
arXiv Detail & Related papers (2020-11-20T19:10:01Z) - Learning Stabilizing Controllers for Unstable Linear Quadratic
Regulators from a Single Trajectory [85.29718245299341]
We study linear controllers under quadratic costs model also known as linear quadratic regulators (LQR)
We present two different semi-definite programs (SDP) which results in a controller that stabilizes all systems within an ellipsoid uncertainty set.
We propose an efficient data dependent algorithm -- textsceXploration -- that with high probability quickly identifies a stabilizing controller.
arXiv Detail & Related papers (2020-06-19T08:58:57Z) - Active Learning for Nonlinear System Identification with Guarantees [102.43355665393067]
We study a class of nonlinear dynamical systems whose state transitions depend linearly on a known feature embedding of state-action pairs.
We propose an active learning approach that achieves this by repeating three steps: trajectory planning, trajectory tracking, and re-estimation of the system from all available data.
We show that our method estimates nonlinear dynamical systems at a parametric rate, similar to the statistical rate of standard linear regression.
arXiv Detail & Related papers (2020-06-18T04:54:11Z) - Learning Constrained Adaptive Differentiable Predictive Control Policies
With Guarantees [1.1086440815804224]
We present differentiable predictive control (DPC), a method for learning constrained neural control policies for linear systems.
We employ automatic differentiation to obtain direct policy gradients by backpropagating the model predictive control (MPC) loss function and constraints penalties through a differentiable closed-loop system dynamics model.
arXiv Detail & Related papers (2020-04-23T14:24:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.