Data Driven Control with Learned Dynamics: Model-Based versus Model-Free
Approach
- URL: http://arxiv.org/abs/2006.09543v1
- Date: Tue, 16 Jun 2020 22:18:21 GMT
- Title: Data Driven Control with Learned Dynamics: Model-Based versus Model-Free
Approach
- Authors: Wenjian Hao, Yiqiang Han
- Abstract summary: We compare two types of data-driven control methods, representing model-based and model-free approaches.
One is a recently proposed method - Deep Koopman Representation for Control (DKRC), which utilizes a deep neural network to map an unknown nonlinear dynamical system to a high-dimensional linear system.
The other is a classic model-free control method based on an actor-critic architecture - Deep Deterministic Policy Gradient (DDPG), which has been proved to be effective in various dynamical systems.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper compares two different types of data-driven control methods,
representing model-based and model-free approaches. One is a recently proposed
method - Deep Koopman Representation for Control (DKRC), which utilizes a deep
neural network to map an unknown nonlinear dynamical system to a
high-dimensional linear system, which allows for employing state-of-the-art
control strategy. The other one is a classic model-free control method based on
an actor-critic architecture - Deep Deterministic Policy Gradient (DDPG), which
has been proved to be effective in various dynamical systems. The comparison is
carried out in OpenAI Gym, which provides multiple control environments for
benchmark purposes. Two examples are provided for comparison, i.e., classic
Inverted Pendulum and Lunar Lander Continuous Control. From the results of the
experiments, we compare these two methods in terms of control strategies and
the effectiveness under various initialization conditions. We also examine the
learned dynamic model from DKRC with the analytical model derived from the
Euler-Lagrange Linearization method, which demonstrates the accuracy in the
learned model for unknown dynamics from a data-driven sample-efficient
approach.
Related papers
- Active Learning for Control-Oriented Identification of Nonlinear Systems [26.231260751633307]
We present the first finite sample analysis of an active learning algorithm suitable for a general class of nonlinear dynamics.
In certain settings, the excess control cost of our algorithm achieves the optimal rate, up to logarithmic factors.
We validate our approach in simulation, showcasing the advantage of active, control-oriented exploration for controlling nonlinear systems.
arXiv Detail & Related papers (2024-04-13T15:40:39Z) - Data-driven Nonlinear Model Reduction using Koopman Theory: Integrated
Control Form and NMPC Case Study [56.283944756315066]
We propose generic model structures combining delay-coordinate encoding of measurements and full-state decoding to integrate reduced Koopman modeling and state estimation.
A case study demonstrates that our approach provides accurate control models and enables real-time capable nonlinear model predictive control of a high-purity cryogenic distillation column.
arXiv Detail & Related papers (2024-01-09T11:54:54Z) - A Data-Driven Model-Reference Adaptive Control Approach Based on
Reinforcement Learning [4.817429789586126]
A model-reference adaptive solution is developed here for autonomous systems where it solves the Hamilton-Jacobi-Bellman equation of an error-based structure.
This is done in real-time without knowing or employing the dynamics of either the process or reference model in the control strategies.
arXiv Detail & Related papers (2023-03-17T14:10:52Z) - Model-Based Reinforcement Learning with SINDy [0.0]
We propose a novel method for discovering the governing non-linear dynamics of physical systems in reinforcement learning (RL)
We establish that this method is capable of discovering the underlying dynamics using significantly fewer trajectories than state of the art model learning algorithms.
arXiv Detail & Related papers (2022-08-30T19:03:48Z) - Dynamically-Scaled Deep Canonical Correlation Analysis [77.34726150561087]
Canonical Correlation Analysis (CCA) is a method for feature extraction of two views by finding maximally correlated linear projections of them.
We introduce a novel dynamic scaling method for training an input-dependent canonical correlation model.
arXiv Detail & Related papers (2022-03-23T12:52:49Z) - Learning to Generate Content-Aware Dynamic Detectors [62.74209921174237]
We introduce a newpective of designing efficient detectors, which is automatically generating sample-adaptive model architecture.
We introduce a course-to-fine strat-egy tailored for object detection to guide the learning of dynamic routing.
Experiments on MS-COCO dataset demonstrate that CADDet achieves 1.8 higher mAP with 10% fewer FLOPs compared with vanilla routing.
arXiv Detail & Related papers (2020-12-08T08:05:20Z) - DyNODE: Neural Ordinary Differential Equations for Dynamics Modeling in
Continuous Control [0.0]
We present a novel approach that captures the underlying dynamics of a system by incorporating control in a neural ordinary differential equation framework.
Results indicate that a simple DyNODE architecture when combined with an actor-critic reinforcement learning algorithm outperforms canonical neural networks.
arXiv Detail & Related papers (2020-09-09T12:56:58Z) - Control as Hybrid Inference [62.997667081978825]
We present an implementation of CHI which naturally mediates the balance between iterative and amortised inference.
We verify the scalability of our algorithm on a continuous control benchmark, demonstrating that it outperforms strong model-free and model-based baselines.
arXiv Detail & Related papers (2020-07-11T19:44:09Z) - Combining Model-Based and Model-Free Methods for Nonlinear Control: A
Provably Convergent Policy Gradient Approach [10.648049177775686]
We develop a novel approach to use the linear model to define a warm start for a model-free, policy method.
We show this hybrid approach outperforms the model-based controller while avoiding the convergence issues associated with model-free approaches.
arXiv Detail & Related papers (2020-06-12T21:16:29Z) - Data-driven Koopman Operators for Model-based Shared Control of
Human-Machine Systems [66.65503164312705]
We present a data-driven shared control algorithm that can be used to improve a human operator's control of complex machines.
Both the dynamics and information about the user's interaction are learned from observation through the use of a Koopman operator.
We find that model-based shared control significantly improves task and control metrics when compared to a natural learning, or user only, control paradigm.
arXiv Detail & Related papers (2020-06-12T14:14:07Z) - Model-Augmented Actor-Critic: Backpropagating through Paths [81.86992776864729]
Current model-based reinforcement learning approaches use the model simply as a learned black-box simulator.
We show how to make more effective use of the model by exploiting its differentiability.
arXiv Detail & Related papers (2020-05-16T19:18:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.