Linear quadratic control of nonlinear systems with Koopman operator
learning and the Nystr\"om method
- URL: http://arxiv.org/abs/2403.02811v1
- Date: Tue, 5 Mar 2024 09:28:40 GMT
- Title: Linear quadratic control of nonlinear systems with Koopman operator
learning and the Nystr\"om method
- Authors: Edoardo Caldarelli, Antoine Chatalic, Adri\`a Colom\'e, Cesare
Molinari, Carlos Ocampo-Martinez, Carme Torras, Lorenzo Rosasco
- Abstract summary: We show how random subspaces can be used to achieve huge computational savings.
Our main technical contribution is deriving theoretical guarantees on the effect of the Nystr"om approximation.
- Score: 15.747820715709937
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this paper, we study how the Koopman operator framework can be combined
with kernel methods to effectively control nonlinear dynamical systems. While
kernel methods have typically large computational requirements, we show how
random subspaces (Nystr\"om approximation) can be used to achieve huge
computational savings while preserving accuracy. Our main technical
contribution is deriving theoretical guarantees on the effect of the Nystr\"om
approximation. More precisely, we study the linear quadratic regulator problem,
showing that both the approximated Riccati operator and the regulator
objective, for the associated solution of the optimal control problem, converge
at the rate $m^{-1/2}$, where $m$ is the random subspace size. Theoretical
findings are complemented by numerical experiments corroborating our results.
Related papers
- Enhanced Feature Learning via Regularisation: Integrating Neural Networks and Kernel Methods [0.0]
We introduce an efficient method for the estimator, called Brownian Kernel Neural Network (BKerNN)
We show that BKerNN's expected risk converges to the minimal risk with explicit high-probability rates of $O( min((d/n)1/2, n-1/6)$ (up to logarithmic factors)
arXiv Detail & Related papers (2024-07-24T13:46:50Z) - Adaptive Federated Learning Over the Air [108.62635460744109]
We propose a federated version of adaptive gradient methods, particularly AdaGrad and Adam, within the framework of over-the-air model training.
Our analysis shows that the AdaGrad-based training algorithm converges to a stationary point at the rate of $mathcalO( ln(T) / T 1 - frac1alpha ).
arXiv Detail & Related papers (2024-03-11T09:10:37Z) - Impact of Computation in Integral Reinforcement Learning for
Continuous-Time Control [5.126167270246931]
We show that the choice of the computational method -- in this case, the quadrature rule -- can significantly impact control performance.
We draw a parallel between IntRL's policy iteration and Newton's method applied to the Hamilton-Jacobi-Bellman equation.
We prove that the local convergence rates for IntRL using the trapezoidal rule and Bayesian quadrature with a Mat'ern kernel to be $O(N-2)$ and $O(N-b)$.
arXiv Detail & Related papers (2024-02-27T10:12:47Z) - A Specialized Semismooth Newton Method for Kernel-Based Optimal
Transport [92.96250725599958]
Kernel-based optimal transport (OT) estimators offer an alternative, functional estimation procedure to address OT problems from samples.
We show that our SSN method achieves a global convergence rate of $O (1/sqrtk)$, and a local quadratic convergence rate under standard regularity conditions.
arXiv Detail & Related papers (2023-10-21T18:48:45Z) - Equation Discovery with Bayesian Spike-and-Slab Priors and Efficient Kernels [57.46832672991433]
We propose a novel equation discovery method based on Kernel learning and BAyesian Spike-and-Slab priors (KBASS)
We use kernel regression to estimate the target function, which is flexible, expressive, and more robust to data sparsity and noises.
We develop an expectation-propagation expectation-maximization algorithm for efficient posterior inference and function estimation.
arXiv Detail & Related papers (2023-10-09T03:55:09Z) - Deep Learning with Kernels through RKHM and the Perron-Frobenius
Operator [14.877070496733966]
Reproducing kernel Hilbert $C*$-module (RKHM) is a generalization of reproducing kernel Hilbert space (RKHS) by means of $C*$-algebra.
We derive a new Rademacher generalization bound in this setting and provide a theoretical interpretation of benign overfitting by means of Perron-Frobenius operators.
arXiv Detail & Related papers (2023-05-23T01:38:41Z) - Learning "best" kernels from data in Gaussian process regression. With
application to aerodynamics [0.4588028371034406]
We introduce algorithms to select/design kernels in Gaussian process regression/kriging surrogate modeling techniques.
A first class of algorithms is kernel flow, which was introduced in a context of classification in machine learning.
A second class of algorithms is called spectral kernel ridge regression, and aims at selecting a "best" kernel such that the norm of the function to be approximated is minimal.
arXiv Detail & Related papers (2022-06-03T07:50:54Z) - Nystr\"om Kernel Mean Embeddings [92.10208929236826]
We propose an efficient approximation procedure based on the Nystr"om method.
It yields sufficient conditions on the subsample size to obtain the standard $n-1/2$ rate.
We discuss applications of this result for the approximation of the maximum mean discrepancy and quadrature rules.
arXiv Detail & Related papers (2022-01-31T08:26:06Z) - Deep Learning Approximation of Diffeomorphisms via Linear-Control
Systems [91.3755431537592]
We consider a control system of the form $dot x = sum_i=1lF_i(x)u_i$, with linear dependence in the controls.
We use the corresponding flow to approximate the action of a diffeomorphism on a compact ensemble of points.
arXiv Detail & Related papers (2021-10-24T08:57:46Z) - Control Occupation Kernel Regression for Nonlinear Control-Affine
Systems [6.308539010172309]
This manuscript presents an algorithm for obtaining an approximation of nonlinear high order control affine dynamical systems.
The vector valued structure of the Hilbert space allows for simultaneous approximation of the drift and control effectiveness components of the control affine system.
arXiv Detail & Related papers (2021-05-31T21:14:30Z) - Adaptive Control and Regret Minimization in Linear Quadratic Gaussian
(LQG) Setting [91.43582419264763]
We propose LqgOpt, a novel reinforcement learning algorithm based on the principle of optimism in the face of uncertainty.
LqgOpt efficiently explores the system dynamics, estimates the model parameters up to their confidence interval, and deploys the controller of the most optimistic model.
arXiv Detail & Related papers (2020-03-12T19:56:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.