Optimisation of Structured Neural Controller Based on Continuous-Time
Policy Gradient
- URL: http://arxiv.org/abs/2201.06262v1
- Date: Mon, 17 Jan 2022 08:06:19 GMT
- Title: Optimisation of Structured Neural Controller Based on Continuous-Time
Policy Gradient
- Authors: Namhoon Cho, Hyo-Sang Shin
- Abstract summary: This study presents a policy optimisation framework for structured nonlinear control of continuous-time (deterministic) dynamic systems.
The proposed approach prescribes a structure for the controller based on relevant scientific knowledge.
Numerical experiments on aerospace applications illustrate the utility of the structured nonlinear controller optimisation framework.
- Score: 2.297079626504224
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This study presents a policy optimisation framework for structured nonlinear
control of continuous-time (deterministic) dynamic systems. The proposed
approach prescribes a structure for the controller based on relevant scientific
knowledge (such as Lyapunov stability theory or domain experiences) while
considering the tunable elements inside the given structure as the point of
parametrisation with neural networks. To optimise a cost represented as a
function of the neural network weights, the proposed approach utilises the
continuous-time policy gradient method based on adjoint sensitivity analysis as
a means for correct and performant computation of cost gradient. This enables
combining the stability, robustness, and physical interpretability of an
analytically-derived structure for the feedback controller with the
representational flexibility and optimised resulting performance provided by
machine learning techniques. Such a hybrid paradigm for fixed-structure control
synthesis is particularly useful for optimising adaptive nonlinear controllers
to achieve improved performance in online operation, an area where the existing
theory prevails the design of structure while lacking clear analytical
understandings about tuning of the gains and the uncertainty model basis
functions that govern the performance characteristics. Numerical experiments on
aerospace applications illustrate the utility of the structured nonlinear
controller optimisation framework.
Related papers
- Receding Hamiltonian-Informed Optimal Neural Control and State Estimation for Closed-Loop Dynamical Systems [4.05766189327054]
Hamiltonian-Informed Optimal Neural (Hion) controllers are a novel class of neural network-based controllers for dynamical systems.
Hion controllers estimate future states and compute optimal control inputs using Pontryagin's Principle.
arXiv Detail & Related papers (2024-11-02T16:06:29Z) - Parameter-Efficient Fine-Tuning for Continual Learning: A Neural Tangent Kernel Perspective [125.00228936051657]
We introduce NTK-CL, a novel framework that eliminates task-specific parameter storage while adaptively generating task-relevant features.
By fine-tuning optimizable parameters with appropriate regularization, NTK-CL achieves state-of-the-art performance on established PEFT-CL benchmarks.
arXiv Detail & Related papers (2024-07-24T09:30:04Z) - Learning to Boost the Performance of Stable Nonlinear Systems [0.0]
We tackle the performance-boosting problem with closed-loop stability guarantees.
Our methods enable learning over arbitrarily deep neural network classes of performance-boosting controllers for stable nonlinear systems.
arXiv Detail & Related papers (2024-05-01T21:11:29Z) - Continual Model-based Reinforcement Learning for Data Efficient Wireless Network Optimisation [73.04087903322237]
We formulate throughput optimisation as Continual Reinforcement Learning of control policies.
Simulation results suggest that the proposed system is able to shorten the end-to-end deployment lead-time by two-fold.
arXiv Detail & Related papers (2024-04-30T11:23:31Z) - A Passivity-Based Method for Accelerated Convex Optimisation [0.8747606955991705]
This study presents a constructive methodology for designing accelerated convex optimisation algorithms in continuous-time domain.
The two key enablers are the classical concept of passivity in control theory and the time-dependent change of variables.
arXiv Detail & Related papers (2023-06-20T11:54:36Z) - Sparsity in Partially Controllable Linear Systems [56.142264865866636]
We study partially controllable linear dynamical systems specified by an underlying sparsity pattern.
Our results characterize those state variables which are irrelevant for optimal control.
arXiv Detail & Related papers (2021-10-12T16:41:47Z) - A Theoretical Overview of Neural Contraction Metrics for Learning-based
Control with Guaranteed Stability [7.963506386866862]
This paper presents a neural network model of an optimal contraction metric and corresponding differential Lyapunov function.
Its innovation lies in providing formal robustness guarantees for learning-based control frameworks.
arXiv Detail & Related papers (2021-10-02T00:28:49Z) - Probabilistic robust linear quadratic regulators with Gaussian processes [73.0364959221845]
Probabilistic models such as Gaussian processes (GPs) are powerful tools to learn unknown dynamical systems from data for subsequent use in control design.
We present a novel controller synthesis for linearized GP dynamics that yields robust controllers with respect to a probabilistic stability margin.
arXiv Detail & Related papers (2021-05-17T08:36:18Z) - COMBO: Conservative Offline Model-Based Policy Optimization [120.55713363569845]
Uncertainty estimation with complex models, such as deep neural networks, can be difficult and unreliable.
We develop a new model-based offline RL algorithm, COMBO, that regularizes the value function on out-of-support state-actions.
We find that COMBO consistently performs as well or better as compared to prior offline model-free and model-based methods.
arXiv Detail & Related papers (2021-02-16T18:50:32Z) - Gaussian Process-based Min-norm Stabilizing Controller for
Control-Affine Systems with Uncertain Input Effects and Dynamics [90.81186513537777]
We propose a novel compound kernel that captures the control-affine nature of the problem.
We show that this resulting optimization problem is convex, and we call it Gaussian Process-based Control Lyapunov Function Second-Order Cone Program (GP-CLF-SOCP)
arXiv Detail & Related papers (2020-11-14T01:27:32Z) - Reinforcement Learning of Structured Control for Linear Systems with
Unknown State Matrix [0.0]
We bring forth the ideas from reinforcement learning (RL) in conjunction with sufficient stability and performance guarantees.
A special control structure enabled by this RL framework is distributed learning control which is necessary for many large-scale cyber-physical systems.
arXiv Detail & Related papers (2020-11-02T17:04:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.