Recurrent Neural Networks for Learning Long-term Temporal Dependencies
with Reanalysis of Time Scale Representation
- URL: http://arxiv.org/abs/2111.03282v1
- Date: Fri, 5 Nov 2021 06:22:58 GMT
- Title: Recurrent Neural Networks for Learning Long-term Temporal Dependencies
with Reanalysis of Time Scale Representation
- Authors: Kentaro Ohno, Atsutoshi Kumagai
- Abstract summary: We argue that the interpretation of a forget gate as a temporal representation is valid when the gradient of loss with respect to the state decreases exponentially as time goes back.
We propose an approach to construct new RNNs that can represent a longer time scale than conventional models.
- Score: 16.32068729107421
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Recurrent neural networks with a gating mechanism such as an LSTM or GRU are
powerful tools to model sequential data. In the mechanism, a forget gate, which
was introduced to control information flow in a hidden state in the RNN, has
recently been re-interpreted as a representative of the time scale of the
state, i.e., a measure how long the RNN retains information on inputs. On the
basis of this interpretation, several parameter initialization methods to
exploit prior knowledge on temporal dependencies in data have been proposed to
improve learnability. However, the interpretation relies on various unrealistic
assumptions, such as that there are no inputs after a certain time point. In
this work, we reconsider this interpretation of the forget gate in a more
realistic setting. We first generalize the existing theory on gated RNNs so
that we can consider the case where inputs are successively given. We then
argue that the interpretation of a forget gate as a temporal representation is
valid when the gradient of loss with respect to the state decreases
exponentially as time goes back. We empirically demonstrate that existing RNNs
satisfy this gradient condition at the initial training phase on several tasks,
which is in good agreement with previous initialization methods. On the basis
of this finding, we propose an approach to construct new RNNs that can
represent a longer time scale than conventional models, which will improve the
learnability for long-term sequential data. We verify the effectiveness of our
method by experiments with real-world datasets.
Related papers
- Deep Neural Networks Tend To Extrapolate Predictably [51.303814412294514]
neural network predictions tend to be unpredictable and overconfident when faced with out-of-distribution (OOD) inputs.
We observe that neural network predictions often tend towards a constant value as input data becomes increasingly OOD.
We show how one can leverage our insights in practice to enable risk-sensitive decision-making in the presence of OOD inputs.
arXiv Detail & Related papers (2023-10-02T03:25:32Z) - Hidden State Approximation in Recurrent Neural Networks Using Continuous
Particle Filtering [0.0]
Using historical data to predict future events has many applications in the real world, such as stock price prediction; the robot localization.
In this paper, we use the particles to approximate the distribution of the latent state and show how it can extend into a more complex form.
With the proposed continuous differentiable scheme, our model is capable of adaptively extracting valuable information and updating the latent state according to the Bayes rule.
arXiv Detail & Related papers (2022-12-18T04:31:45Z) - Learning Low Dimensional State Spaces with Overparameterized Recurrent
Neural Nets [57.06026574261203]
We provide theoretical evidence for learning low-dimensional state spaces, which can also model long-term memory.
Experiments corroborate our theory, demonstrating extrapolation via learning low-dimensional state spaces with both linear and non-linear RNNs.
arXiv Detail & Related papers (2022-10-25T14:45:15Z) - Reducing Catastrophic Forgetting in Self Organizing Maps with
Internally-Induced Generative Replay [67.50637511633212]
A lifelong learning agent is able to continually learn from potentially infinite streams of pattern sensory data.
One major historic difficulty in building agents that adapt is that neural systems struggle to retain previously-acquired knowledge when learning from new samples.
This problem is known as catastrophic forgetting (interference) and remains an unsolved problem in the domain of machine learning to this day.
arXiv Detail & Related papers (2021-12-09T07:11:14Z) - Spike-inspired Rank Coding for Fast and Accurate Recurrent Neural
Networks [5.986408771459261]
Biological spiking neural networks (SNNs) can temporally encode information in their outputs, whereas artificial neural networks (ANNs) conventionally do not.
Here we show that temporal coding such as rank coding (RC) inspired by SNNs can also be applied to conventional ANNs such as LSTMs.
RC-training also significantly reduces time-to-insight during inference, with a minimal decrease in accuracy.
We demonstrate these in two toy problems of sequence classification, and in a temporally-encoded MNIST dataset where our RC model achieves 99.19% accuracy after the first input time-step
arXiv Detail & Related papers (2021-10-06T15:51:38Z) - Training Feedback Spiking Neural Networks by Implicit Differentiation on
the Equilibrium State [66.2457134675891]
Spiking neural networks (SNNs) are brain-inspired models that enable energy-efficient implementation on neuromorphic hardware.
Most existing methods imitate the backpropagation framework and feedforward architectures for artificial neural networks.
We propose a novel training method that does not rely on the exact reverse of the forward computation.
arXiv Detail & Related papers (2021-09-29T07:46:54Z) - Online learning of windmill time series using Long Short-term Cognitive
Networks [58.675240242609064]
The amount of data generated on windmill farms makes online learning the most viable strategy to follow.
We use Long Short-term Cognitive Networks (LSTCNs) to forecast windmill time series in online settings.
Our approach reported the lowest forecasting errors with respect to a simple RNN, a Long Short-term Memory, a Gated Recurrent Unit, and a Hidden Markov Model.
arXiv Detail & Related papers (2021-07-01T13:13:24Z) - UnICORNN: A recurrent model for learning very long time dependencies [0.0]
We propose a novel RNN architecture based on a structure preserving discretization of a Hamiltonian system of second-order ordinary differential equations.
The resulting RNN is fast, invertible (in time), memory efficient and we derive rigorous bounds on the hidden state gradients to prove the mitigation of the exploding and vanishing gradient problem.
arXiv Detail & Related papers (2021-03-09T15:19:59Z) - Multi-Time-Scale Input Approaches for Hourly-Scale Rainfall-Runoff
Modeling based on Recurrent Neural Networks [0.0]
Two approaches are proposed to reduce the required computational time for time-series modeling through a recurrent neural network (RNN)
One approach provides coarse fine temporal resolutions of the input time-series to RNN in parallel.
The results confirm that both of the proposed approaches can reduce the computational time for the training of RNN significantly.
arXiv Detail & Related papers (2021-01-30T07:51:55Z) - Frequentist Uncertainty in Recurrent Neural Networks via Blockwise
Influence Functions [121.10450359856242]
Recurrent neural networks (RNNs) are instrumental in modelling sequential and time-series data.
Existing approaches for uncertainty quantification in RNNs are based predominantly on Bayesian methods.
We develop a frequentist alternative that: (a) does not interfere with model training or compromise its accuracy, (b) applies to any RNN architecture, and (c) provides theoretical coverage guarantees on the estimated uncertainty intervals.
arXiv Detail & Related papers (2020-06-20T22:45:32Z) - Depth Enables Long-Term Memory for Recurrent Neural Networks [0.0]
We introduce a measure of the network's ability to support information flow across time, referred to as the Start-End separation rank.
We prove that deep recurrent networks support Start-End separation ranks which are higher than those supported by their shallow counterparts.
arXiv Detail & Related papers (2020-03-23T10:29:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.