Accelerated Continuous-Time Approximate Dynamic Programming via
Data-Assisted Hybrid Control
- URL: http://arxiv.org/abs/2204.12707v1
- Date: Wed, 27 Apr 2022 05:36:51 GMT
- Title: Accelerated Continuous-Time Approximate Dynamic Programming via
Data-Assisted Hybrid Control
- Authors: Daniel E. Ochoa, Jorge I. Poveda
- Abstract summary: We introduce an algorithm that incorporates dynamic momentum in actor-critic structures to control continuous-time dynamic plants with an affine structure in the input.
By incorporating dynamic momentum in our algorithm, we are able to accelerate the convergence properties of the closed-loop system.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We introduce a new closed-loop architecture for the online solution of
approximate optimal control problems in the context of continuous-time systems.
Specifically, we introduce the first algorithm that incorporates dynamic
momentum in actor-critic structures to control continuous-time dynamic plants
with an affine structure in the input. By incorporating dynamic momentum in our
algorithm, we are able to accelerate the convergence properties of the
closed-loop system, achieving superior transient performance compared to
traditional gradient-descent based techniques. In addition, by leveraging the
existence of past recorded data with sufficiently rich information properties,
we dispense with the persistence of excitation condition traditionally imposed
on the regressors of the critic and the actor. Given that our continuous-time
momentum-based dynamics also incorporate periodic discrete-time resets that
emulate restarting techniques used in the machine learning literature, we
leverage tools from hybrid dynamical systems theory to establish asymptotic
stability properties for the closed-loop system. We illustrate our results with
a numerical example.
Related papers
- Amortized Control of Continuous State Space Feynman-Kac Model for Irregular Time Series [14.400596021890863]
Many real-world datasets, such as healthcare, climate, and economics, are often collected as irregular time series.
We propose the Amortized Control of continuous State Space Model (ACSSM) for continuous dynamical modeling of time series.
arXiv Detail & Related papers (2024-10-08T01:27:46Z) - Learning Noise-Robust Stable Koopman Operator for Control with Hankel DMD [1.0742675209112622]
We propose a noise-robust learning framework for the Koopman operator of nonlinear dynamical systems.
We develop a stable parameterization of the Koopman operator, along with a progressive learning strategy for roll-out recurrent loss.
arXiv Detail & Related papers (2024-08-13T03:39:34Z) - Physics-Inspired Temporal Learning of Quadrotor Dynamics for Accurate
Model Predictive Trajectory Tracking [76.27433308688592]
Accurately modeling quadrotor's system dynamics is critical for guaranteeing agile, safe, and stable navigation.
We present a novel Physics-Inspired Temporal Convolutional Network (PI-TCN) approach to learning quadrotor's system dynamics purely from robot experience.
Our approach combines the expressive power of sparse temporal convolutions and dense feed-forward connections to make accurate system predictions.
arXiv Detail & Related papers (2022-06-07T13:51:35Z) - Model-Based Reinforcement Learning via Stochastic Hybrid Models [39.83837705993256]
This paper adopts a hybrid-system view of nonlinear modeling and control.
We consider a sequence modeling paradigm that captures the temporal structure of the data.
We show that these time-series models naturally admit a closed-loop extension that we use to extract local feedback controllers.
arXiv Detail & Related papers (2021-11-11T14:05:46Z) - Deep Explicit Duration Switching Models for Time Series [84.33678003781908]
We propose a flexible model that is capable of identifying both state- and time-dependent switching dynamics.
State-dependent switching is enabled by a recurrent state-to-switch connection.
An explicit duration count variable is used to improve the time-dependent switching behavior.
arXiv Detail & Related papers (2021-10-26T17:35:21Z) - Value Iteration in Continuous Actions, States and Time [99.00362538261972]
We propose a continuous fitted value iteration (cFVI) algorithm for continuous states and actions.
The optimal policy can be derived for non-linear control-affine dynamics.
Videos of the physical system are available at urlhttps://sites.google.com/view/value-iteration.
arXiv Detail & Related papers (2021-05-10T21:40:56Z) - Stochastically forced ensemble dynamic mode decomposition for
forecasting and analysis of near-periodic systems [65.44033635330604]
We introduce a novel load forecasting method in which observed dynamics are modeled as a forced linear system.
We show that its use of intrinsic linear dynamics offers a number of desirable properties in terms of interpretability and parsimony.
Results are presented for a test case using load data from an electrical grid.
arXiv Detail & Related papers (2020-10-08T20:25:52Z) - Online Reinforcement Learning Control by Direct Heuristic Dynamic
Programming: from Time-Driven to Event-Driven [80.94390916562179]
Time-driven learning refers to the machine learning method that updates parameters in a prediction model continuously as new data arrives.
It is desirable to prevent the time-driven dHDP from updating due to insignificant system event such as noise.
We show how the event-driven dHDP algorithm works in comparison to the original time-driven dHDP.
arXiv Detail & Related papers (2020-06-16T05:51:25Z) - Liquid Time-constant Networks [117.57116214802504]
We introduce a new class of time-continuous recurrent neural network models.
Instead of declaring a learning system's dynamics by implicit nonlinearities, we construct networks of linear first-order dynamical systems.
These neural networks exhibit stable and bounded behavior, yield superior expressivity within the family of neural ordinary differential equations.
arXiv Detail & Related papers (2020-06-08T09:53:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.