Hybrid Backpropagation Parallel Reservoir Networks
- URL: http://arxiv.org/abs/2010.14611v1
- Date: Tue, 27 Oct 2020 21:03:35 GMT
- Title: Hybrid Backpropagation Parallel Reservoir Networks
- Authors: Matthew Evanusa and Snehesh Shrestha and Michelle Girvan and Cornelia
Ferm\"uller and Yiannis Aloimonos
- Abstract summary: We propose a novel hybrid network, which combines the effectiveness of learning random temporal features of reservoirs with the readout power of a deep neural network with batch normalization.
We demonstrate that our new network outperforms LSTMs and GRUs, including multi-layer "deep" versions of these networks.
We show also that the inclusion of a novel meta-ring structure, which we call HBP-ESN M-Ring, achieves similar performance to one large reservoir while decreasing the memory required by an order of magnitude.
- Score: 8.944918753413827
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In many real-world applications, fully-differentiable RNNs such as LSTMs and
GRUs have been widely deployed to solve time series learning tasks. These
networks train via Backpropagation Through Time, which can work well in
practice but involves a biologically unrealistic unrolling of the network in
time for gradient updates, are computationally expensive, and can be hard to
tune. A second paradigm, Reservoir Computing, keeps the recurrent weight matrix
fixed and random. Here, we propose a novel hybrid network, which we call Hybrid
Backpropagation Parallel Echo State Network (HBP-ESN) which combines the
effectiveness of learning random temporal features of reservoirs with the
readout power of a deep neural network with batch normalization. We demonstrate
that our new network outperforms LSTMs and GRUs, including multi-layer "deep"
versions of these networks, on two complex real-world multi-dimensional time
series datasets: gesture recognition using skeleton keypoints from ChaLearn,
and the DEAP dataset for emotion recognition from EEG measurements. We show
also that the inclusion of a novel meta-ring structure, which we call HBP-ESN
M-Ring, achieves similar performance to one large reservoir while decreasing
the memory required by an order of magnitude. We thus offer this new hybrid
reservoir deep learning paradigm as a new alternative direction for RNN
learning of temporal or sequential data.
Related papers
- Properties and Potential Applications of Random Functional-Linked Types
of Neural Networks [81.56822938033119]
Random functional-linked neural networks (RFLNNs) offer an alternative way of learning in deep structure.
This paper gives some insights into the properties of RFLNNs from the viewpoints of frequency domain.
We propose a method to generate a BLS network with better performance, and design an efficient algorithm for solving Poison's equation.
arXiv Detail & Related papers (2023-04-03T13:25:22Z) - Learning Fast and Slow for Online Time Series Forecasting [76.50127663309604]
Fast and Slow learning Networks (FSNet) is a holistic framework for online time-series forecasting.
FSNet balances fast adaptation to recent changes and retrieving similar old knowledge.
Our code will be made publicly available.
arXiv Detail & Related papers (2022-02-23T18:23:07Z) - Parallel Machine Learning for Forecasting the Dynamics of Complex
Networks [0.0]
We present a machine learning scheme for forecasting the dynamics of large complex networks.
We use a parallel architecture that mimics the topology of the network of interest.
arXiv Detail & Related papers (2021-08-27T06:06:41Z) - Online learning of windmill time series using Long Short-term Cognitive
Networks [58.675240242609064]
The amount of data generated on windmill farms makes online learning the most viable strategy to follow.
We use Long Short-term Cognitive Networks (LSTCNs) to forecast windmill time series in online settings.
Our approach reported the lowest forecasting errors with respect to a simple RNN, a Long Short-term Memory, a Gated Recurrent Unit, and a Hidden Markov Model.
arXiv Detail & Related papers (2021-07-01T13:13:24Z) - Revisiting the double-well problem by deep learning with a hybrid
network [7.308730248177914]
We propose a novel hybrid network which integrates two different kinds of neural networks: LSTM and ResNet.
Such a hybrid network can be applied for solving cooperative dynamics in a system with fast spatial or temporal modulations.
arXiv Detail & Related papers (2021-04-25T07:51:43Z) - Contextual HyperNetworks for Novel Feature Adaptation [43.49619456740745]
Contextual HyperNetwork (CHN) generates parameters for extending the base model to a new feature.
At prediction time, the CHN requires only a single forward pass through a neural network, yielding a significant speed-up.
We show that this system obtains improved few-shot learning performance for novel features over existing imputation and meta-learning baselines.
arXiv Detail & Related papers (2021-04-12T23:19:49Z) - PredRNN: A Recurrent Neural Network for Spatiotemporal Predictive
Learning [109.84770951839289]
We present PredRNN, a new recurrent network for learning visual dynamics from historical context.
We show that our approach obtains highly competitive results on three standard datasets.
arXiv Detail & Related papers (2021-03-17T08:28:30Z) - Deep Cellular Recurrent Network for Efficient Analysis of Time-Series
Data with Spatial Information [52.635997570873194]
This work proposes a novel deep cellular recurrent neural network (DCRNN) architecture to process complex multi-dimensional time series data with spatial information.
The proposed architecture achieves state-of-the-art performance while utilizing substantially less trainable parameters when compared to comparable methods in the literature.
arXiv Detail & Related papers (2021-01-12T20:08:18Z) - SRDCNN: Strongly Regularized Deep Convolution Neural Network
Architecture for Time-series Sensor Signal Classification Tasks [4.950427992960756]
We present SRDCNN: Strongly Regularized Deep Convolution Neural Network (DCNN) based deep architecture to perform time series classification tasks.
The novelty of the proposed approach is that the network weights are regularized by both L1 and L2 norm penalties.
arXiv Detail & Related papers (2020-07-14T08:42:39Z) - Depth Enables Long-Term Memory for Recurrent Neural Networks [0.0]
We introduce a measure of the network's ability to support information flow across time, referred to as the Start-End separation rank.
We prove that deep recurrent networks support Start-End separation ranks which are higher than those supported by their shallow counterparts.
arXiv Detail & Related papers (2020-03-23T10:29:14Z) - Large-Scale Gradient-Free Deep Learning with Recursive Local
Representation Alignment [84.57874289554839]
Training deep neural networks on large-scale datasets requires significant hardware resources.
Backpropagation, the workhorse for training these networks, is an inherently sequential process that is difficult to parallelize.
We propose a neuro-biologically-plausible alternative to backprop that can be used to train deep networks.
arXiv Detail & Related papers (2020-02-10T16:20:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.