Lyapunov Function-guided Reinforcement Learning for Flight Control
- URL: http://arxiv.org/abs/2510.22840v1
- Date: Sun, 26 Oct 2025 21:18:34 GMT
- Title: Lyapunov Function-guided Reinforcement Learning for Flight Control
- Authors: Yifei Li, Erik-Jan van Kampen,
- Abstract summary: A cascaded online learning flight control system has been developed and enhanced with respect to action smoothness.<n>We investigate the convergence performance of the control system, characterized by the increment of a Lyapunov function candidate.
- Score: 4.349671197115746
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: A cascaded online learning flight control system has been developed and enhanced with respect to action smoothness. In this paper, we investigate the convergence performance of the control system, characterized by the increment of a Lyapunov function candidate. The derivation of this metric accounts for discretization errors and state prediction errors introduced by the incremental model. Comparative results are presented through flight control simulations.
Related papers
- Model-Free versus Model-Based Reinforcement Learning for Fixed-Wing UAV
Attitude Control Under Varying Wind Conditions [1.474723404975345]
This paper evaluates and compares the performance of model-free and model-based reinforcement learning for the attitude control of fixed-wing unmanned aerial vehicles using PID as a reference point.
Results show that the Temporal Difference Model Predictive Control agent outperforms both the PID controller and other model-free reinforcement learning methods in terms of tracking accuracy and robustness.
arXiv Detail & Related papers (2024-09-26T14:47:14Z) - Cooperative Learning with Gaussian Processes for Euler-Lagrange Systems
Tracking Control under Switching Topologies [9.838373797093245]
This work presents an innovative learning-based approach to tackle the tracking control problem of Euler-Lagrange multi-agent systems.
A standout feature is its exceptional efficiency in deriving the aggregation weights achieved.
Simulation experiments validate the protocol's efficacy in effectively managing complex scenarios.
arXiv Detail & Related papers (2024-02-05T14:33:52Z) - Improving the Performance of Robust Control through Event-Triggered
Learning [74.57758188038375]
We propose an event-triggered learning algorithm that decides when to learn in the face of uncertainty in the LQR problem.
We demonstrate improved performance over a robust controller baseline in a numerical example.
arXiv Detail & Related papers (2022-07-28T17:36:37Z) - Lyapunov Density Models: Constraining Distribution Shift in
Learning-Based Control [64.61499213110334]
We seek a mechanism to constrain the agent to states and actions that resemble those that it was trained on.
In control theory, Lyapunov stability and control-invariant sets allow us to make guarantees about controllers.
density models allow us to estimate the training data distribution.
arXiv Detail & Related papers (2022-06-21T16:49:09Z) - Interpretable Stochastic Model Predictive Control using Distributional
Reinforced Estimation for Quadrotor Tracking Systems [0.8411385346896411]
We present a novel trajectory tracker for autonomous quadrotor navigation in dynamic and complex environments.
The proposed framework integrates a distributional Reinforcement Learning estimator for unknown aerodynamic effects into a Model Predictive Controller.
We demonstrate our system to improve the cumulative tracking errors by at least 66% with unknown and diverse aerodynamic forces.
arXiv Detail & Related papers (2022-05-14T23:27:38Z) - Robustness of Control Design via Bayesian Learning [0.0]
Inspired by these findings, we demonstrate the robustness properties of Bayesian learning in the control search task.
We seek to find a linear controller that stabilizes a one-dimensional open-loop unstable system.
arXiv Detail & Related papers (2022-05-13T21:10:19Z) - The Impact of Data on the Stability of Learning-Based Control- Extended
Version [63.97366815968177]
We propose a Lyapunov-based measure for quantifying the impact of data on the certifiable control performance.
By modeling unknown system dynamics through Gaussian processes, we can determine the interrelation between model uncertainty and satisfaction of stability conditions.
arXiv Detail & Related papers (2020-11-20T19:10:01Z) - Gaussian Process-based Min-norm Stabilizing Controller for
Control-Affine Systems with Uncertain Input Effects and Dynamics [90.81186513537777]
We propose a novel compound kernel that captures the control-affine nature of the problem.
We show that this resulting optimization problem is convex, and we call it Gaussian Process-based Control Lyapunov Function Second-Order Cone Program (GP-CLF-SOCP)
arXiv Detail & Related papers (2020-11-14T01:27:32Z) - Certainty Equivalent Perception-Based Control [29.216967322052785]
We show a uniform error bound on non kernel regression under a dynamically-achievable dense sampling scheme.
This allows for a finite-time convergence rate on the sub-optimality of using the regressor in closed-loop for waypoint tracking.
arXiv Detail & Related papers (2020-08-27T18:45:40Z) - Anticipating the Long-Term Effect of Online Learning in Control [75.6527644813815]
AntLer is a design algorithm for learning-based control laws that anticipates learning.
We show that AntLer approximates an optimal solution arbitrarily accurately with probability one.
arXiv Detail & Related papers (2020-07-24T07:00:14Z) - Adaptive Control and Regret Minimization in Linear Quadratic Gaussian
(LQG) Setting [91.43582419264763]
We propose LqgOpt, a novel reinforcement learning algorithm based on the principle of optimism in the face of uncertainty.
LqgOpt efficiently explores the system dynamics, estimates the model parameters up to their confidence interval, and deploys the controller of the most optimistic model.
arXiv Detail & Related papers (2020-03-12T19:56:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.