Sub-optimality of the Separation Principle for Quadratic Control from Bilinear Observations
- URL: http://arxiv.org/abs/2504.11555v1
- Date: Tue, 15 Apr 2025 18:53:51 GMT
- Title: Sub-optimality of the Separation Principle for Quadratic Control from Bilinear Observations
- Authors: Yahya Sattar, Sunmook Choi, Yassir Jedra, Maryam Fazel, Sarah Dean,
- Abstract summary: We consider the problem of controlling a numerical dynamical system from bilinear observations with minimal quadratic cost.<n>Despite the similarity of this problem to standard linear quadratic Gaussian (LQG) filters, we show that when neither is bilinear the Separation Principle model.<n>We illustrate our results through experiments in multiple synthetic settings.
- Score: 16.7109402673221
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We consider the problem of controlling a linear dynamical system from bilinear observations with minimal quadratic cost. Despite the similarity of this problem to standard linear quadratic Gaussian (LQG) control, we show that when the observation model is bilinear, neither does the Separation Principle hold, nor is the optimal controller affine in the estimated state. Moreover, the cost-to-go is non-convex in the control input. Hence, finding an analytical expression for the optimal feedback controller is difficult in general. Under certain settings, we show that the standard LQG controller locally maximizes the cost instead of minimizing it. Furthermore, the optimal controllers (derived analytically) are not unique and are nonlinear in the estimated state. We also introduce a notion of input-dependent observability and derive conditions under which the Kalman filter covariance remains bounded. We illustrate our theoretical results through numerical experiments in multiple synthetic settings.
Related papers
- Ensemble Kalman-Bucy filtering for nonlinear model predictive control [0.32634122554914]
We extend the popular ensemble Kalman filter to receding horizon optimal control problems in the spirit of nonlinear model predictive control.<n>We provide an interacting particle approximation to the forward-backward differential equations arising from Pontryagin's maximum principle.<n>The receding horizon control laws are approximated as linear and are continuously updated as in nonlinear model predictive control.
arXiv Detail & Related papers (2025-03-16T12:04:28Z) - Physics-informed Gaussian Processes as Linear Model Predictive Controller [5.89889361990138]
We introduce a novel algorithm for controlling linear time invariant systems in a tracking problem.<n>The controller is based on a Gaussian Process (GP) whose realizations satisfy a system of linear ordinary differential equations with constant coefficients.
arXiv Detail & Related papers (2024-12-02T15:37:37Z) - Multirotor Ensemble Model Predictive Control I: Simulation Experiments [0.0]
An ensemble-represented Gaussian process performs the backward calculations to determine optimal gains for the initial time.
We construct the EMPC for terminal control and regulation problems and apply it to the control of a simulated, identical-twin study.
arXiv Detail & Related papers (2023-05-22T01:32:17Z) - On the Optimization Landscape of Dynamic Output Feedback: A Case Study
for Linear Quadratic Regulator [12.255864026960403]
We show how the dLQR cost varies with the coordinate transformation of the dynamic controller and then derive the optimal transformation for a given observable stabilizing controller.
These results shed light on designing efficient algorithms for general decision-making problems with partially observed information.
arXiv Detail & Related papers (2022-09-12T06:43:35Z) - Deep Learning Approximation of Diffeomorphisms via Linear-Control
Systems [91.3755431537592]
We consider a control system of the form $dot x = sum_i=1lF_i(x)u_i$, with linear dependence in the controls.
We use the corresponding flow to approximate the action of a diffeomorphism on a compact ensemble of points.
arXiv Detail & Related papers (2021-10-24T08:57:46Z) - Sparsity in Partially Controllable Linear Systems [56.142264865866636]
We study partially controllable linear dynamical systems specified by an underlying sparsity pattern.
Our results characterize those state variables which are irrelevant for optimal control.
arXiv Detail & Related papers (2021-10-12T16:41:47Z) - Robust Online Control with Model Misspecification [96.23493624553998]
We study online control of an unknown nonlinear dynamical system with model misspecification.
Our study focuses on robustness, which measures how much deviation from the assumed linear approximation can be tolerated.
arXiv Detail & Related papers (2021-07-16T07:04:35Z) - Advancing Trajectory Optimization with Approximate Inference:
Exploration, Covariance Control and Adaptive Risk [29.811633555275666]
We look at the input inference for control (i2c) algorithm, and derive three key characteristics that enable advanced trajectory optimization.
An expert' linear Gaussian controller that combines the benefits of open-loop optima and closed-loop variance reduction when optimizing for nonlinear systems.
arXiv Detail & Related papers (2021-03-10T19:52:31Z) - Gaussian Process-based Min-norm Stabilizing Controller for
Control-Affine Systems with Uncertain Input Effects and Dynamics [90.81186513537777]
We propose a novel compound kernel that captures the control-affine nature of the problem.
We show that this resulting optimization problem is convex, and we call it Gaussian Process-based Control Lyapunov Function Second-Order Cone Program (GP-CLF-SOCP)
arXiv Detail & Related papers (2020-11-14T01:27:32Z) - Adaptive Control and Regret Minimization in Linear Quadratic Gaussian
(LQG) Setting [91.43582419264763]
We propose LqgOpt, a novel reinforcement learning algorithm based on the principle of optimism in the face of uncertainty.
LqgOpt efficiently explores the system dynamics, estimates the model parameters up to their confidence interval, and deploys the controller of the most optimistic model.
arXiv Detail & Related papers (2020-03-12T19:56:38Z) - Improper Learning for Non-Stochastic Control [78.65807250350755]
We consider the problem of controlling a possibly unknown linear dynamical system with adversarial perturbations, adversarially chosen convex loss functions, and partially observed states.
Applying online descent to this parametrization yields a new controller which attains sublinear regret vs. a large class of closed-loop policies.
Our bounds are the first in the non-stochastic control setting that compete with emphall stabilizing linear dynamical controllers.
arXiv Detail & Related papers (2020-01-25T02:12:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.