On the Convexity of Discrete Time Covariance Steering in Stochastic
Linear Systems with Wasserstein Terminal Cost
- URL: http://arxiv.org/abs/2103.13579v1
- Date: Thu, 25 Mar 2021 03:24:52 GMT
- Title: On the Convexity of Discrete Time Covariance Steering in Stochastic
Linear Systems with Wasserstein Terminal Cost
- Authors: Isin M. Balci, Abhishek Halder, Efstathios Bakolas
- Abstract summary: We show that when the terminal state covariance is upper bounded, with respect to the L"owner partial order, our problem admits a unique global minimizing state feedback gain.
The results of this paper set the stage for the development of specialized control design tools.
- Score: 1.1602089225841632
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this work, we analyze the properties of the solution to the covariance
steering problem for discrete time Gaussian linear systems with a squared
Wasserstein distance terminal cost. In our previous work, we have shown that by
utilizing the state feedback control policy parametrization, this stochastic
optimal control problem can be associated with a difference of convex functions
program. Here, we revisit the same covariance control problem but this time we
focus on the analysis of the problem. Specifically, we establish the existence
of solutions to the optimization problem and derive the first and second order
conditions for optimality. We provide analytic expressions for the gradient and
the Hessian of the performance index by utilizing specialized tools from matrix
calculus. Subsequently, we prove that the optimization problem always admits a
global minimizer, and finally, we provide a sufficient condition for the
performance index to be a strictly convex function (under the latter condition,
the problem admits a unique global minimizer). In particular, we show that when
the terminal state covariance is upper bounded, with respect to the L\"{o}wner
partial order, by the covariance matrix of the desired terminal normal
distribution, then our problem admits a unique global minimizing state feedback
gain. The results of this paper set the stage for the development of
specialized control design tools that exploit the structure of the solution to
the covariance steering problem with a squared Wasserstein distance terminal
cost.
Related papers
- Error Feedback under $(L_0,L_1)$-Smoothness: Normalization and Momentum [56.37522020675243]
We provide the first proof of convergence for normalized error feedback algorithms across a wide range of machine learning problems.
We show that due to their larger allowable stepsizes, our new normalized error feedback algorithms outperform their non-normalized counterparts on various tasks.
arXiv Detail & Related papers (2024-10-22T10:19:27Z) - Harmonic Path Integral Diffusion [0.4527270266697462]
We present a novel approach for sampling from a continuous multivariate probability distribution, which may either be explicitly known (up to a normalization factor) or represented via empirical samples.
Our method constructs a time-dependent bridge from a delta function centered at the origin of the state space at $t=0$, transforming it into the target distribution at $t=1$.
We contrast these algorithms with other sampling methods, particularly simulated and path integral sampling, highlighting their advantages in terms of analytical control, accuracy, and computational efficiency.
arXiv Detail & Related papers (2024-09-23T16:20:21Z) - Constrained Optimization via Exact Augmented Lagrangian and Randomized
Iterative Sketching [55.28394191394675]
We develop an adaptive inexact Newton method for equality-constrained nonlinear, nonIBS optimization problems.
We demonstrate the superior performance of our method on benchmark nonlinear problems, constrained logistic regression with data from LVM, and a PDE-constrained problem.
arXiv Detail & Related papers (2023-05-28T06:33:37Z) - Fully Stochastic Trust-Region Sequential Quadratic Programming for
Equality-Constrained Optimization Problems [62.83783246648714]
We propose a sequential quadratic programming algorithm (TR-StoSQP) to solve nonlinear optimization problems with objectives and deterministic equality constraints.
The algorithm adaptively selects the trust-region radius and, compared to the existing line-search StoSQP schemes, allows us to utilize indefinite Hessian matrices.
arXiv Detail & Related papers (2022-11-29T05:52:17Z) - A Physics-informed Deep Learning Approach for Minimum Effort Stochastic
Control of Colloidal Self-Assembly [9.791617215182598]
The control objective is formulated in terms of steering the state PDFs from a prescribed initial probability measure towards a prescribed terminal probability measure with minimum control effort.
We derive the conditions of optimality for the associated optimal control problem.
The performance of the proposed solution is demonstrated via numerical simulations on a benchmark colloidal self-assembly problem.
arXiv Detail & Related papers (2022-08-19T07:01:57Z) - Optimal Rates for Random Order Online Optimization [60.011653053877126]
We study the citetgarber 2020online, where the loss functions may be chosen by an adversary, but are then presented online in a uniformly random order.
We show that citetgarber 2020online algorithms achieve the optimal bounds and significantly improve their stability.
arXiv Detail & Related papers (2021-06-29T09:48:46Z) - Gaussian Process-based Min-norm Stabilizing Controller for
Control-Affine Systems with Uncertain Input Effects and Dynamics [90.81186513537777]
We propose a novel compound kernel that captures the control-affine nature of the problem.
We show that this resulting optimization problem is convex, and we call it Gaussian Process-based Control Lyapunov Function Second-Order Cone Program (GP-CLF-SOCP)
arXiv Detail & Related papers (2020-11-14T01:27:32Z) - An improved convergence analysis for decentralized online stochastic
non-convex optimization [17.386715847732468]
In this paper, we show that a technique called GT-Loakjasiewics (GT-Loakjasiewics) satisfies the existing condition GT-Loakjasiewics (GT-Loakjasiewics) satisfies the current best convergence rates.
The results are not only immediately applicable but also the currently known best convergence rates.
arXiv Detail & Related papers (2020-08-10T15:29:13Z) - Provably Convergent Working Set Algorithm for Non-Convex Regularized
Regression [0.0]
This paper proposes a working set algorithm for non-regular regularizers with convergence guarantees.
Our results demonstrate high gain compared to the full problem solver for both block-coordinates or a gradient solver.
arXiv Detail & Related papers (2020-06-24T07:40:31Z) - Convergence of adaptive algorithms for weakly convex constrained
optimization [59.36386973876765]
We prove the $mathcaltilde O(t-1/4)$ rate of convergence for the norm of the gradient of Moreau envelope.
Our analysis works with mini-batch size of $1$, constant first and second order moment parameters, and possibly smooth optimization domains.
arXiv Detail & Related papers (2020-06-11T17:43:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.