A Data-Driven Model-Reference Adaptive Control Approach Based on
Reinforcement Learning
- URL: http://arxiv.org/abs/2303.09994v1
- Date: Fri, 17 Mar 2023 14:10:52 GMT
- Title: A Data-Driven Model-Reference Adaptive Control Approach Based on
Reinforcement Learning
- Authors: Mohammed Abouheaf, Wail Gueaieb, Davide Spinello and Salah Al-Sharhan
- Abstract summary: A model-reference adaptive solution is developed here for autonomous systems where it solves the Hamilton-Jacobi-Bellman equation of an error-based structure.
This is done in real-time without knowing or employing the dynamics of either the process or reference model in the control strategies.
- Score: 4.817429789586126
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Model-reference adaptive systems refer to a consortium of techniques that
guide plants to track desired reference trajectories. Approaches based on
theories like Lyapunov, sliding surfaces, and backstepping are typically
employed to advise adaptive control strategies. The resulting solutions are
often challenged by the complexity of the reference model and those of the
derived control strategies. Additionally, the explicit dependence of the
control strategies on the process dynamics and reference dynamical models may
contribute in degrading their efficiency in the face of uncertain or unknown
dynamics. A model-reference adaptive solution is developed here for autonomous
systems where it solves the Hamilton-Jacobi-Bellman equation of an error-based
structure. The proposed approach describes the process with an integral
temporal difference equation and solves it using an integral reinforcement
learning mechanism. This is done in real-time without knowing or employing the
dynamics of either the process or reference model in the control strategies. A
class of aircraft is adopted to validate the proposed technique.
Related papers
- Data-driven Nonlinear Model Reduction using Koopman Theory: Integrated
Control Form and NMPC Case Study [56.283944756315066]
We propose generic model structures combining delay-coordinate encoding of measurements and full-state decoding to integrate reduced Koopman modeling and state estimation.
A case study demonstrates that our approach provides accurate control models and enables real-time capable nonlinear model predictive control of a high-purity cryogenic distillation column.
arXiv Detail & Related papers (2024-01-09T11:54:54Z) - An Adaptive Fuzzy Reinforcement Learning Cooperative Approach for the
Autonomous Control of Flock Systems [4.961066282705832]
This work introduces an adaptive distributed robustness technique for the autonomous control of flock systems.
Its relatively flexible structure is based on online fuzzy reinforcement learning schemes which simultaneously target a number of objectives.
In addition to its resilience in the face of dynamic disturbances, the algorithm does not require more than the agent position as a feedback signal.
arXiv Detail & Related papers (2023-03-17T13:07:35Z) - When to Update Your Model: Constrained Model-based Reinforcement
Learning [50.74369835934703]
We propose a novel and general theoretical scheme for a non-decreasing performance guarantee of model-based RL (MBRL)
Our follow-up derived bounds reveal the relationship between model shifts and performance improvement.
A further example demonstrates that learning models from a dynamically-varying number of explorations benefit the eventual returns.
arXiv Detail & Related papers (2022-10-15T17:57:43Z) - Planning with Diffusion for Flexible Behavior Synthesis [125.24438991142573]
We consider what it would look like to fold as much of the trajectory optimization pipeline as possible into the modeling problem.
The core of our technical approach lies in a diffusion probabilistic model that plans by iteratively denoising trajectories.
arXiv Detail & Related papers (2022-05-20T07:02:03Z) - Structured Hammerstein-Wiener Model Learning for Model Predictive
Control [0.2752817022620644]
This paper aims to improve the reliability of optimal control using models constructed by machine learning methods.
In this paper, we propose a model that combines the Hammerstein-Wiener model with convex neural networks.
arXiv Detail & Related papers (2021-07-09T06:41:34Z) - Trajectory Tracking of Underactuated Sea Vessels With Uncertain
Dynamics: An Integral Reinforcement Learning Approach [2.064612766965483]
An online machine learning mechanism based on integral reinforcement learning is proposed to find a solution for a class of nonlinear tracking problems.
The solution is implemented using an online value iteration process which is realized by employing means of the adaptive critics and gradient descent approaches.
arXiv Detail & Related papers (2021-04-01T01:41:49Z) - Control as Hybrid Inference [62.997667081978825]
We present an implementation of CHI which naturally mediates the balance between iterative and amortised inference.
We verify the scalability of our algorithm on a continuous control benchmark, demonstrating that it outperforms strong model-free and model-based baselines.
arXiv Detail & Related papers (2020-07-11T19:44:09Z) - Data Driven Control with Learned Dynamics: Model-Based versus Model-Free
Approach [0.0]
We compare two types of data-driven control methods, representing model-based and model-free approaches.
One is a recently proposed method - Deep Koopman Representation for Control (DKRC), which utilizes a deep neural network to map an unknown nonlinear dynamical system to a high-dimensional linear system.
The other is a classic model-free control method based on an actor-critic architecture - Deep Deterministic Policy Gradient (DDPG), which has been proved to be effective in various dynamical systems.
arXiv Detail & Related papers (2020-06-16T22:18:21Z) - Localized active learning of Gaussian process state space models [63.97366815968177]
A globally accurate model is not required to achieve good performance in many common control applications.
We propose an active learning strategy for Gaussian process state space models that aims to obtain an accurate model on a bounded subset of the state-action space.
By employing model predictive control, the proposed technique integrates information collected during exploration and adaptively improves its exploration strategy.
arXiv Detail & Related papers (2020-05-04T05:35:02Z) - Information Theoretic Model Predictive Q-Learning [64.74041985237105]
We present a novel theoretical connection between information theoretic MPC and entropy regularized RL.
We develop a Q-learning algorithm that can leverage biased models.
arXiv Detail & Related papers (2019-12-31T00:29:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.