Control of a fly-mimicking flyer in complex flow using deep
reinforcement learning
- URL: http://arxiv.org/abs/2111.03454v1
- Date: Thu, 4 Nov 2021 04:48:56 GMT
- Title: Control of a fly-mimicking flyer in complex flow using deep
reinforcement learning
- Authors: Seungpyo Hong, Sejin Kim, Donghyun You
- Abstract summary: An integrated framework of computational fluid-structural dynamics (CFD-CSD) and deep reinforcement learning (deep-RL) is developed for control of a fly-scale flexible-winged flyer in complex flow.
To obtain accurate data, the CFD-CSD is adopted for precisely predicting the dynamics.
To gain ample data, a novel data reproduction method is devised, where the obtained data are replicated for various situations.
- Score: 0.12891210250935145
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: An integrated framework of computational fluid-structural dynamics (CFD-CSD)
and deep reinforcement learning (deep-RL) is developed for control of a
fly-scale flexible-winged flyer in complex flow. Dynamics of the flyer in
complex flow is highly unsteady and nonlinear, which makes modeling the
dynamics challenging. Thus, conventional control methodologies, where the
dynamics is modeled, are insufficient for regulating such complicated dynamics.
Therefore, in the present study, the integrated framework, in which the whole
governing equations for fluid and structure are solved, is proposed to generate
a control policy for the flyer. For the deep-RL to successfully learn the
control policy, accurate and ample data of the dynamics are required. However,
satisfying both the quality and quantity of the data on the intricate dynamics
is extremely difficult since, in general, more accurate data are more costly.
In the present study, two strategies are proposed to deal with the dilemma. To
obtain accurate data, the CFD-CSD is adopted for precisely predicting the
dynamics. To gain ample data, a novel data reproduction method is devised,
where the obtained data are replicated for various situations while conserving
the dynamics. With those data, the framework learns the control policy in
various flow conditions and the learned policy is shown to have remarkable
performance in controlling the flyer in complex flow fields.
Related papers
- SINDy-RL: Interpretable and Efficient Model-Based Reinforcement Learning [5.59265003686955]
We introduce SINDy-RL, a framework for combining SINDy and deep reinforcement learning.
SINDy-RL achieves comparable performance to state-of-the-art DRL algorithms.
We demonstrate the effectiveness of our approaches on benchmark control environments and challenging fluids problems.
arXiv Detail & Related papers (2024-03-14T05:17:39Z) - Controlling dynamical systems to complex target states using machine
learning: next-generation vs. classical reservoir computing [68.8204255655161]
Controlling nonlinear dynamical systems using machine learning allows to drive systems into simple behavior like periodicity but also to more complex arbitrary dynamics.
We show first that classical reservoir computing excels at this task.
In a next step, we compare those results based on different amounts of training data to an alternative setup, where next-generation reservoir computing is used instead.
It turns out that while delivering comparable performance for usual amounts of training data, next-generation RC significantly outperforms in situations where only very limited data is available.
arXiv Detail & Related papers (2023-07-14T07:05:17Z) - How to Control Hydrodynamic Force on Fluidic Pinball via Deep
Reinforcement Learning [3.1635451288803638]
We present a DRL-based real-time feedback strategy to control the hydrodynamic force on fluidic pinball.
By adequately designing reward functions and encoding historical observations, the DRL-based control was shown to make reasonable and valid control decisions.
One of these results was analyzed by a machine learning model that enabled us to shed light on the basis of decision-making and physical mechanisms of the force tracking process.
arXiv Detail & Related papers (2023-04-23T03:39:50Z) - Learning Control-Oriented Dynamical Structure from Data [25.316358215670274]
We discuss a state-dependent nonlinear tracking controller formulation for general nonlinear control-affine systems.
We empirically demonstrate the efficacy of learned versions of this controller in stable trajectory tracking.
arXiv Detail & Related papers (2023-02-06T02:01:38Z) - Data efficient reinforcement learning and adaptive optimal perimeter
control of network traffic dynamics [0.0]
This work proposes an integral reinforcement learning (IRL) based approach to learning the macroscopic traffic dynamics for adaptive optimal perimeter control.
To reduce the sampling complexity and use the available data more efficiently, the experience replay (ER) technique is introduced to the IRL algorithm.
The convergence of the IRL-based algorithms and the stability of the controlled traffic dynamics are proven via the Lyapunov theory.
arXiv Detail & Related papers (2022-09-13T04:28:49Z) - Physics-Inspired Temporal Learning of Quadrotor Dynamics for Accurate
Model Predictive Trajectory Tracking [76.27433308688592]
Accurately modeling quadrotor's system dynamics is critical for guaranteeing agile, safe, and stable navigation.
We present a novel Physics-Inspired Temporal Convolutional Network (PI-TCN) approach to learning quadrotor's system dynamics purely from robot experience.
Our approach combines the expressive power of sparse temporal convolutions and dense feed-forward connections to make accurate system predictions.
arXiv Detail & Related papers (2022-06-07T13:51:35Z) - Capturing Actionable Dynamics with Structured Latent Ordinary
Differential Equations [68.62843292346813]
We propose a structured latent ODE model that captures system input variations within its latent representation.
Building on a static variable specification, our model learns factors of variation for each input to the system, thus separating the effects of the system inputs in the latent space.
arXiv Detail & Related papers (2022-02-25T20:00:56Z) - Multi-intersection Traffic Optimisation: A Benchmark Dataset and a
Strong Baseline [85.9210953301628]
Control of traffic signals is fundamental and critical to alleviate traffic congestion in urban areas.
Because of the high complexity of modelling the problem, experimental settings of current works are often inconsistent.
We propose a novel and strong baseline model based on deep reinforcement learning with the encoder-decoder structure.
arXiv Detail & Related papers (2021-01-24T03:55:39Z) - Context-aware Dynamics Model for Generalization in Model-Based
Reinforcement Learning [124.9856253431878]
We decompose the task of learning a global dynamics model into two stages: (a) learning a context latent vector that captures the local dynamics, then (b) predicting the next state conditioned on it.
In order to encode dynamics-specific information into the context latent vector, we introduce a novel loss function that encourages the context latent vector to be useful for predicting both forward and backward dynamics.
The proposed method achieves superior generalization ability across various simulated robotics and control tasks, compared to existing RL schemes.
arXiv Detail & Related papers (2020-05-14T08:10:54Z) - Hierarchical Decomposition of Nonlinear Dynamics and Control for System
Identification and Policy Distillation [39.83837705993256]
Current trends in reinforcement learning (RL) focus on complex representations of dynamics and policies.
We take inspiration from the control community and apply the principles of hybrid switching systems in order to break down complex dynamics into simpler components.
arXiv Detail & Related papers (2020-05-04T12:40:59Z) - Information Theoretic Model Predictive Q-Learning [64.74041985237105]
We present a novel theoretical connection between information theoretic MPC and entropy regularized RL.
We develop a Q-learning algorithm that can leverage biased models.
arXiv Detail & Related papers (2019-12-31T00:29:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.