Deep polytopic autoencoders for low-dimensional linear parameter-varying approximations and nonlinear feedback design
- URL: http://arxiv.org/abs/2403.18044v2
- Date: Thu, 23 Jan 2025 02:28:53 GMT
- Title: Deep polytopic autoencoders for low-dimensional linear parameter-varying approximations and nonlinear feedback design
- Authors: Jan Heiland, Yongho Kim, Steffen W. R. Werner,
- Abstract summary: Polytopic autoencoders provide low-di-men-sion-al parametrizations of states in a polytope.
For nonlinear PDEs, this is readily applied to low-dimensional linear parameter-varying (LPV) approximations.
- Score: 0.9187159782788578
- License:
- Abstract: Polytopic autoencoders provide low-di\-men\-sion\-al parametrizations of states in a polytope. For nonlinear PDEs, this is readily applied to low-dimensional linear parameter-varying (LPV) approximations as they have been exploited for efficient nonlinear controller design via series expansions of the solution to the state-dependent Riccati equation. In this work, we develop a polytopic autoencoder for control applications and show how it improves on standard linear approaches in view of LPV approximations of nonlinear systems. We discuss how the particular architecture enables exact representation of target states and higher order series expansions of the nonlinear feedback law at little extra computational effort in the online phase and how the linear though high-dimensional and nonstandard Lyapunov equations are efficiently computed during the offline phase. In a numerical study, we illustrate the procedure and how this approach can reliably outperform the standard linear-quadratic regulator design.
Related papers
- Estimation Sample Complexity of a Class of Nonlinear Continuous-time Systems [0.0]
We present a method of parameter estimation for large class of nonlinear systems, namely those in which the state consists of output derivatives and the flow is linear in the parameter.
The method, which solves for the unknown parameter by directly inverting the dynamics using regularized linear regression, is based on new design and analysis ideas for differentiation filtering and regularized least squares.
arXiv Detail & Related papers (2023-12-08T21:42:11Z) - Pessimistic Nonlinear Least-Squares Value Iteration for Offline Reinforcement Learning [53.97335841137496]
We propose an oracle-efficient algorithm, dubbed Pessimistic Least-Square Value Iteration (PNLSVI) for offline RL with non-linear function approximation.
Our algorithm enjoys a regret bound that has a tight dependency on the function class complexity and achieves minimax optimal instance-dependent regret when specialized to linear function approximation.
arXiv Detail & Related papers (2023-10-02T17:42:01Z) - Discrete-Time Nonlinear Feedback Linearization via Physics-Informed
Machine Learning [0.0]
We present a physics-informed machine learning scheme for the feedback linearization of nonlinear systems.
We show that the proposed PIML outperforms the traditional numerical implementation.
arXiv Detail & Related papers (2023-03-15T19:03:23Z) - Convolutional Autoencoders, Clustering and POD for Low-dimensional
Parametrization of Navier-Stokes Equations [1.160208922584163]
We propose a convolutional autoencoder (CAE) consisting of a nonlinear encoder and an affine linear decoder.
The proposed set of methods is compared to the standard POD approach in two cylinder-wake scenarios modeled by the incompressible Navier-Stokes equations.
arXiv Detail & Related papers (2023-02-02T18:12:08Z) - Linear Convergence of Natural Policy Gradient Methods with Log-Linear
Policies [115.86431674214282]
We consider infinite-horizon discounted Markov decision processes and study the convergence rates of the natural policy gradient (NPG) and the Q-NPG methods with the log-linear policy class.
We show that both methods attain linear convergence rates and $mathcalO (1/epsilon2)$ sample complexities using a simple, non-adaptive geometrically increasing step size.
arXiv Detail & Related papers (2022-10-04T06:17:52Z) - Sample Efficient Reinforcement Learning In Continuous State Spaces: A
Perspective Beyond Linearity [50.38337893712897]
We introduce the Effective Planning Window (EPW) condition, a structural condition on MDPs that makes no linearity assumptions.
We demonstrate that the EPW condition permits sample efficient RL, by providing an algorithm which provably solves MDPs satisfying this condition.
We additionally show the necessity of conditions like EPW, by demonstrating that simple MDPs with slight nonlinearities cannot be solved sample efficiently.
arXiv Detail & Related papers (2021-06-15T00:06:59Z) - Sample-Efficient Reinforcement Learning Is Feasible for Linearly
Realizable MDPs with Limited Revisiting [60.98700344526674]
Low-complexity models such as linear function representation play a pivotal role in enabling sample-efficient reinforcement learning.
In this paper, we investigate a new sampling protocol, which draws samples in an online/exploratory fashion but allows one to backtrack and revisit previous states in a controlled and infrequent manner.
We develop an algorithm tailored to this setting, achieving a sample complexity that scales practicallyly with the feature dimension, the horizon, and the inverse sub-optimality gap, but not the size of the state/action space.
arXiv Detail & Related papers (2021-05-17T17:22:07Z) - POD-DL-ROM: enhancing deep learning-based reduced order models for
nonlinear parametrized PDEs by proper orthogonal decomposition [0.0]
Deep learning-based reduced order models (DL-ROMs) have been recently proposed to overcome common limitations shared by conventional reduced order models (ROMs)
In this paper we propose a possible way to avoid an expensive training stage of DL-ROMs, by (i) performing a prior dimensionality reduction through POD, and (ii) relying on a multi-fidelity pretraining stage.
The proposed POD-DL-ROM is tested on several (both scalar and vector, linear and nonlinear) time-dependent parametrized PDEs.
arXiv Detail & Related papers (2021-01-28T07:34:15Z) - LQF: Linear Quadratic Fine-Tuning [114.3840147070712]
We present the first method for linearizing a pre-trained model that achieves comparable performance to non-linear fine-tuning.
LQF consists of simple modifications to the architecture, loss function and optimization typically used for classification.
arXiv Detail & Related papers (2020-12-21T06:40:20Z) - Pushing the Envelope of Rotation Averaging for Visual SLAM [69.7375052440794]
We propose a novel optimization backbone for visual SLAM systems.
We leverage averaging to improve the accuracy, efficiency and robustness of conventional monocular SLAM systems.
Our approach can exhibit up to 10x faster with comparable accuracy against the state-art on public benchmarks.
arXiv Detail & Related papers (2020-11-02T18:02:26Z) - A comprehensive deep learning-based approach to reduced order modeling
of nonlinear time-dependent parametrized PDEs [0.0]
We show how to construct a DL-ROM for both linear and nonlinear time-dependent parametrized PDEs.
Numerical results indicate that DL-ROMs whose dimension is equal to the intrinsic dimensionality of the PDE solutions manifold are able to approximate the solution of parametrized PDEs.
arXiv Detail & Related papers (2020-01-12T21:18:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.