Observer-Feedback-Feedforward Controller Structures in Reinforcement
Learning
- URL: http://arxiv.org/abs/2304.10276v1
- Date: Thu, 20 Apr 2023 12:59:21 GMT
- Title: Observer-Feedback-Feedforward Controller Structures in Reinforcement
Learning
- Authors: Ruoqi Zhang, Per Mattson, Torbj\"orn Wigren
- Abstract summary: The paper proposes the use of structured neural networks for reinforcement learning based nonlinear adaptive control.
The focus is on partially observable systems, with separate neural networks for the state and feedforward observer and the state feedback and feedforward controller.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The paper proposes the use of structured neural networks for reinforcement
learning based nonlinear adaptive control. The focus is on partially observable
systems, with separate neural networks for the state and feedforward observer
and the state feedback and feedforward controller. The observer dynamics are
modelled by recurrent neural networks while a standard network is used for the
controller. As discussed in the paper, this leads to a separation of the
observer dynamics to the recurrent neural network part, and the state feedback
to the feedback and feedforward network. The structured approach reduces the
computational complexity and gives the reinforcement learning based controller
an {\em understandable} structure as compared to when one single neural network
is used. As shown by simulation the proposed structure has the additional and
main advantage that the training becomes significantly faster. Two ways to
include feedforward structure are presented, one related to state feedback
control and one related to classical feedforward control. The latter method
introduces further structure with a separate recurrent neural network that
processes only the measured disturbance. When evaluated with simulation on a
nonlinear cascaded double tank process, the method with most structure performs
the best, with excellent feedforward disturbance rejection gains.
Related papers
- Verification of Neural Network Control Systems in Continuous Time [1.5695847325697108]
We develop the first verification method for continuously-actuated neural network control systems.
We accomplish this by adding a level of abstraction to model the neural network controller.
We demonstrate the approach's efficacy by applying it to a vision-based autonomous airplane taxiing system.
arXiv Detail & Related papers (2024-05-31T19:39:48Z) - Leveraging Low-Rank and Sparse Recurrent Connectivity for Robust
Closed-Loop Control [63.310780486820796]
We show how a parameterization of recurrent connectivity influences robustness in closed-loop settings.
We find that closed-form continuous-time neural networks (CfCs) with fewer parameters can outperform their full-rank, fully-connected counterparts.
arXiv Detail & Related papers (2023-10-05T21:44:18Z) - How neural networks learn to classify chaotic time series [77.34726150561087]
We study the inner workings of neural networks trained to classify regular-versus-chaotic time series.
We find that the relation between input periodicity and activation periodicity is key for the performance of LKCNN models.
arXiv Detail & Related papers (2023-06-04T08:53:27Z) - Minimizing Control for Credit Assignment with Strong Feedback [65.59995261310529]
Current methods for gradient-based credit assignment in deep neural networks need infinitesimally small feedback signals.
We combine strong feedback influences on neural activity with gradient-based learning and show that this naturally leads to a novel view on neural network optimization.
We show that the use of strong feedback in DFC allows learning forward and feedback connections simultaneously, using a learning rule fully local in space and time.
arXiv Detail & Related papers (2022-04-14T22:06:21Z) - Data-driven emergence of convolutional structure in neural networks [83.4920717252233]
We show how fully-connected neural networks solving a discrimination task can learn a convolutional structure directly from their inputs.
By carefully designing data models, we show that the emergence of this pattern is triggered by the non-Gaussian, higher-order local structure of the inputs.
arXiv Detail & Related papers (2022-02-01T17:11:13Z) - Imbedding Deep Neural Networks [0.0]
Continuous depth neural networks, such as Neural ODEs, have refashioned the understanding of residual neural networks in terms of non-linear vector-valued optimal control problems.
We propose a new approach which explicates the network's depth' as a fundamental variable, thus reducing the problem to a system of forward-facing initial value problems.
arXiv Detail & Related papers (2022-01-31T22:00:41Z) - On the role of feedback in visual processing: a predictive coding
perspective [0.6193838300896449]
We consider deep convolutional networks (CNNs) as models of feed-forward visual processing and implement Predictive Coding (PC) dynamics.
We find that the network increasingly relies on top-down predictions as the noise level increases.
In addition, the accuracy of the network implementing PC dynamics significantly increases over time-steps, compared to its equivalent forward network.
arXiv Detail & Related papers (2021-06-08T10:07:23Z) - Local Critic Training for Model-Parallel Learning of Deep Neural
Networks [94.69202357137452]
We propose a novel model-parallel learning method, called local critic training.
We show that the proposed approach successfully decouples the update process of the layer groups for both convolutional neural networks (CNNs) and recurrent neural networks (RNNs)
We also show that trained networks by the proposed method can be used for structural optimization.
arXiv Detail & Related papers (2021-02-03T09:30:45Z) - Implicit recurrent networks: A novel approach to stationary input
processing with recurrent neural networks in deep learning [0.0]
In this work, we introduce and test a novel implementation of recurrent neural networks into deep learning.
We provide an algorithm which implements the backpropagation algorithm on a implicit implementation of recurrent networks.
A single-layer implicit recurrent network is able to solve the XOR problem, while a feed-forward network with monotonically increasing activation function fails at this task.
arXiv Detail & Related papers (2020-10-20T18:55:32Z) - Neural Networks with Recurrent Generative Feedback [61.90658210112138]
We instantiate this design on convolutional neural networks (CNNs)
In the experiments, CNN-F shows considerably improved adversarial robustness over conventional feedforward CNNs on standard benchmarks.
arXiv Detail & Related papers (2020-07-17T19:32:48Z) - Lifted Regression/Reconstruction Networks [17.89437720094451]
We propose lifted regression/reconstruction networks (LRRNs)
LRRNs combine lifted neural networks with a guaranteed Lipschitz continuity property for the output layer.
We analyse and numerically demonstrate applications for unsupervised and supervised learning.
arXiv Detail & Related papers (2020-05-07T13:24:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.