On the Effectiveness of Iterative Learning Control
- URL: http://arxiv.org/abs/2111.09434v1
- Date: Wed, 17 Nov 2021 22:35:39 GMT
- Title: On the Effectiveness of Iterative Learning Control
- Authors: Anirudh Vemula, Wen Sun, Maxim Likhachev, J. Andrew Bagnell
- Abstract summary: Iterative learning control (ILC) is a powerful technique for high performance tracking in the presence of modeling errors.
There is little prior theoretical work that explains the effectiveness of ILC even in the presence of large modeling errors.
We show that the suboptimality gap, as measured with respect to the optimal LQR controller, for ILC is lower than that for MM by higher order terms.
- Score: 28.76900887141432
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Iterative learning control (ILC) is a powerful technique for high performance
tracking in the presence of modeling errors for optimal control applications.
There is extensive prior work showing its empirical effectiveness in
applications such as chemical reactors, industrial robots and quadcopters.
However, there is little prior theoretical work that explains the effectiveness
of ILC even in the presence of large modeling errors, where optimal control
methods using the misspecified model (MM) often perform poorly. Our work
presents such a theoretical study of the performance of both ILC and MM on
Linear Quadratic Regulator (LQR) problems with unknown transition dynamics. We
show that the suboptimality gap, as measured with respect to the optimal LQR
controller, for ILC is lower than that for MM by higher order terms that become
significant in the regime of high modeling errors. A key part of our analysis
is the perturbation bounds for the discrete Ricatti equation in the finite
horizon setting, where the solution is not a fixed point and requires tracking
the error using recursive bounds. We back our theoretical findings with
empirical experiments on a toy linear dynamical system with an approximate
model, a nonlinear inverted pendulum system with misspecified mass, and a
nonlinear planar quadrotor system in the presence of wind. Experiments show
that ILC outperforms MM significantly, in terms of the cost of computed
trajectories, when modeling errors are high.
Related papers
- ODELoRA: Training Low-Rank Adaptation by Solving Ordinary Differential Equations [54.886931928255564]
Low-rank adaptation (LoRA) has emerged as a widely adopted parameter-efficient fine-tuning method in deep transfer learning.<n>We propose a novel continuous-time optimization dynamic for LoRA factor matrices in the form of an ordinary differential equation (ODE)<n>We show that ODELoRA achieves stable feature learning, a property that is crucial for training deep neural networks at different scales of problem dimensionality.
arXiv Detail & Related papers (2026-02-07T10:19:36Z) - Potential failures of physics-informed machine learning in traffic flow modeling: theoretical and experimental analysis [5.937203351551678]
This study critically examines the performance of physics-informed machine learning (PIML) approaches for traffic flow modeling.<n> Experiments reveal that physical residuals can degrade the performance of LWR- and ARZ-based PIML models.<n>We identify the Courant-Friedrichs-Lewy (CFL) condition as a key indicator of dataset suitability for PIML.
arXiv Detail & Related papers (2025-05-16T17:55:06Z) - RoSTE: An Efficient Quantization-Aware Supervised Fine-Tuning Approach for Large Language Models [53.571195477043496]
We propose an algorithm named Rotated Straight-Through-Estimator (RoSTE)
RoSTE combines quantization-aware supervised fine-tuning (QA-SFT) with an adaptive rotation strategy to reduce activation outliers.
Our findings reveal that the prediction error is directly proportional to the quantization error of the converged weights, which can be effectively managed through an optimized rotation configuration.
arXiv Detail & Related papers (2025-02-13T06:44:33Z) - Koopman-Based Surrogate Modelling of Turbulent Rayleigh-Bénard Convection [4.248022697109535]
We use a Koopman-inspired architecture called the Linear Recurrent Autoencoder Network (LRAN) for learning reduced-order dynamics in convection flows.
A traditional fluid dynamics method, the Kernel Dynamic Mode Decomposition (KDMD) is used to compare the LRAN.
We obtained more accurate predictions with the LRAN than with KDMD in the most turbulent setting.
arXiv Detail & Related papers (2024-05-10T12:15:02Z) - Data-driven Nonlinear Model Reduction using Koopman Theory: Integrated
Control Form and NMPC Case Study [56.283944756315066]
We propose generic model structures combining delay-coordinate encoding of measurements and full-state decoding to integrate reduced Koopman modeling and state estimation.
A case study demonstrates that our approach provides accurate control models and enables real-time capable nonlinear model predictive control of a high-purity cryogenic distillation column.
arXiv Detail & Related papers (2024-01-09T11:54:54Z) - Model-Based Reparameterization Policy Gradient Methods: Theory and
Practical Algorithms [88.74308282658133]
Reization (RP) Policy Gradient Methods (PGMs) have been widely adopted for continuous control tasks in robotics and computer graphics.
Recent studies have revealed that, when applied to long-term reinforcement learning problems, model-based RP PGMs may experience chaotic and non-smooth optimization landscapes.
We propose a spectral normalization method to mitigate the exploding variance issue caused by long model unrolls.
arXiv Detail & Related papers (2023-10-30T18:43:21Z) - Learning Nonlinear Projections for Reduced-Order Modeling of Dynamical
Systems using Constrained Autoencoders [0.0]
We introduce a class of nonlinear projections described by constrained autoencoder neural networks in which both the manifold and the projection fibers are learned from data.
Our architecture uses invertible activation functions and biorthogonal weight matrices to ensure that the encoder is a left inverse of the decoder.
We also introduce new dynamics-aware cost functions that promote learning of oblique projection fibers that account for fast dynamics and nonnormality.
arXiv Detail & Related papers (2023-07-28T04:01:48Z) - Multirotor Ensemble Model Predictive Control I: Simulation Experiments [0.0]
An ensemble-represented Gaussian process performs the backward calculations to determine optimal gains for the initial time.
We construct the EMPC for terminal control and regulation problems and apply it to the control of a simulated, identical-twin study.
arXiv Detail & Related papers (2023-05-22T01:32:17Z) - Two-step reinforcement learning for model-free redesign of nonlinear
optimal regulator [1.5624421399300306]
Reinforcement learning (RL) is one of the promising approaches that enable model-free redesign of optimal controllers for nonlinear dynamical systems.
We propose a model-free two-step design approach that improves the transient learning performance of RL in an optimal regulator redesign problem for unknown nonlinear systems.
arXiv Detail & Related papers (2021-03-05T17:12:33Z) - Gaussian Process-based Min-norm Stabilizing Controller for
Control-Affine Systems with Uncertain Input Effects and Dynamics [90.81186513537777]
We propose a novel compound kernel that captures the control-affine nature of the problem.
We show that this resulting optimization problem is convex, and we call it Gaussian Process-based Control Lyapunov Function Second-Order Cone Program (GP-CLF-SOCP)
arXiv Detail & Related papers (2020-11-14T01:27:32Z) - Derivative-Based Koopman Operators for Real-Time Control of Robotic
Systems [14.211417879279075]
This paper presents a generalizable methodology for data-driven identification of nonlinear dynamics that bounds the model error.
We construct a Koopman operator-based linear representation and utilize Taylor series accuracy analysis to derive an error bound.
When combined with control, the Koopman representation of the nonlinear system has marginally better performance than competing nonlinear modeling methods.
arXiv Detail & Related papers (2020-10-12T15:15:13Z) - Adaptive Control and Regret Minimization in Linear Quadratic Gaussian
(LQG) Setting [91.43582419264763]
We propose LqgOpt, a novel reinforcement learning algorithm based on the principle of optimism in the face of uncertainty.
LqgOpt efficiently explores the system dynamics, estimates the model parameters up to their confidence interval, and deploys the controller of the most optimistic model.
arXiv Detail & Related papers (2020-03-12T19:56:38Z) - Information Theoretic Model Predictive Q-Learning [64.74041985237105]
We present a novel theoretical connection between information theoretic MPC and entropy regularized RL.
We develop a Q-learning algorithm that can leverage biased models.
arXiv Detail & Related papers (2019-12-31T00:29:22Z) - Localized Debiased Machine Learning: Efficient Inference on Quantile
Treatment Effects and Beyond [69.83813153444115]
We consider an efficient estimating equation for the (local) quantile treatment effect ((L)QTE) in causal inference.
Debiased machine learning (DML) is a data-splitting approach to estimating high-dimensional nuisances.
We propose localized debiased machine learning (LDML), which avoids this burdensome step.
arXiv Detail & Related papers (2019-12-30T14:42:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.