On the Stability of Nonlinear Receding Horizon Control: A Geometric
Perspective
- URL: http://arxiv.org/abs/2103.15010v3
- Date: Thu, 25 Jan 2024 19:02:39 GMT
- Title: On the Stability of Nonlinear Receding Horizon Control: A Geometric
Perspective
- Authors: Tyler Westenbroek, Max Simchowitz, Michael I. Jordan, S. Shankar
Sastry
- Abstract summary: widespread adoption of nonlinear Receding Control (RHC) strategies by industry has to more than 30 years.
This paper takes the first step towards understanding the role of global geometry in the role of global-based control.
- Score: 72.7951562665449
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: %!TEX root = LCSS_main_max.tex
The widespread adoption of nonlinear Receding Horizon Control (RHC)
strategies by industry has led to more than 30 years of intense research
efforts to provide stability guarantees for these methods. However, current
theoretical guarantees require that each (generally nonconvex) planning problem
can be solved to (approximate) global optimality, which is an unrealistic
requirement for the derivative-based local optimization methods generally used
in practical implementations of RHC. This paper takes the first step towards
understanding stability guarantees for nonlinear RHC when the inner planning
problem is solved to first-order stationary points, but not necessarily global
optima. Special attention is given to feedback linearizable systems, and a
mixture of positive and negative results are provided. We establish that, under
certain strong conditions, first-order solutions to RHC exponentially stabilize
linearizable systems. Surprisingly, these conditions can hold even in
situations where there may be \textit{spurious local minima.} Crucially, this
guarantee requires that state costs applied to the planning problems are in a
certain sense `compatible' with the global geometry of the system, and a simple
counter-example demonstrates the necessity of this condition. These results
highlight the need to rethink the role of global geometry in the context of
optimization-based control.
Related papers
- Policy Gradient Converges to the Globally Optimal Policy for Nearly Linear-Quadratic Regulators [11.400431211239958]
We study the optimal rate in nearly linear-quadratic regulator systems.
We propose a policy that is guaranteed to the globally optimal rate with a gradient algorithm.
arXiv Detail & Related papers (2023-03-15T08:08:02Z) - Offline Minimax Soft-Q-learning Under Realizability and Partial Coverage [100.8180383245813]
We propose value-based algorithms for offline reinforcement learning (RL)
We show an analogous result for vanilla Q-functions under a soft margin condition.
Our algorithms' loss functions arise from casting the estimation problems as nonlinear convex optimization problems and Lagrangifying.
arXiv Detail & Related papers (2023-02-05T14:22:41Z) - Fully Stochastic Trust-Region Sequential Quadratic Programming for
Equality-Constrained Optimization Problems [62.83783246648714]
We propose a sequential quadratic programming algorithm (TR-StoSQP) to solve nonlinear optimization problems with objectives and deterministic equality constraints.
The algorithm adaptively selects the trust-region radius and, compared to the existing line-search StoSQP schemes, allows us to utilize indefinite Hessian matrices.
arXiv Detail & Related papers (2022-11-29T05:52:17Z) - Dynamic selection of p-norm in linear adaptive filtering via online
kernel-based reinforcement learning [8.319127681936815]
This study addresses the problem of selecting dynamically, at each time instance, the optimal'' p-norm to combat outliers in linear adaptive filtering.
Online and data-driven framework is designed via kernel-based reinforcement learning (KBRL)
arXiv Detail & Related papers (2022-10-20T14:49:39Z) - The Probabilistic Normal Epipolar Constraint for Frame-To-Frame Rotation
Optimization under Uncertain Feature Positions [53.478856119297284]
We introduce the probabilistic normal epipolar constraint (PNEC) that overcomes the limitation by accounting for anisotropic and inhomogeneous uncertainties in the feature positions.
In experiments on synthetic data, we demonstrate that the novel PNEC yields more accurate rotation estimates than the original NEC.
We integrate the proposed method into a state-of-the-art monocular rotation-only odometry system and achieve consistently improved results for the real-world KITTI dataset.
arXiv Detail & Related papers (2022-04-05T14:47:11Z) - Neural System Level Synthesis: Learning over All Stabilizing Policies
for Nonlinear Systems [0.0]
We propose a Neural SLS (Neur-SLS) approach guaranteeing closed-loop stability during and after parameter optimization.
We exploit recent Deep Neural Network (DNN) models based on Recurrent Equilibrium Networks (RENs) to learn over a rich class of nonlinear stable operators.
arXiv Detail & Related papers (2022-03-22T15:22:31Z) - Youla-REN: Learning Nonlinear Feedback Policies with Robust Stability
Guarantees [5.71097144710995]
This paper presents a parameterization of nonlinear controllers for uncertain systems building on a recently developed neural network architecture.
The proposed framework has "built-in" guarantees of stability, i.e., all policies in the search space result in a contracting (globally exponentially stable) closed-loop system.
arXiv Detail & Related papers (2021-12-02T13:52:37Z) - Pointwise Feasibility of Gaussian Process-based Safety-Critical Control
under Model Uncertainty [77.18483084440182]
Control Barrier Functions (CBFs) and Control Lyapunov Functions (CLFs) are popular tools for enforcing safety and stability of a controlled system, respectively.
We present a Gaussian Process (GP)-based approach to tackle the problem of model uncertainty in safety-critical controllers that use CBFs and CLFs.
arXiv Detail & Related papers (2021-06-13T23:08:49Z) - Policy Mirror Descent for Regularized Reinforcement Learning: A
Generalized Framework with Linear Convergence [60.20076757208645]
This paper proposes a general policy mirror descent (GPMD) algorithm for solving regularized RL.
We demonstrate that our algorithm converges linearly over an entire range learning rates, in a dimension-free fashion, to the global solution.
arXiv Detail & Related papers (2021-05-24T02:21:34Z) - Derivative-Free Policy Optimization for Risk-Sensitive and Robust
Control Design: Implicit Regularization and Sample Complexity [15.940861063732608]
Direct policy search serves as one of the workhorses in modern reinforcement learning (RL)
We investigate the convergence theory of policy robustness (PG) methods for the linear risk-sensitive and robust controller.
One feature of our algorithms is that during the learning phase, a certain level complexity/risk-sensitivity controller is preserved.
arXiv Detail & Related papers (2021-01-04T16:00:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.