Tensor-Train Recurrent Neural Networks for Interpretable Multi-Way
Financial Forecasting
- URL: http://arxiv.org/abs/2105.04983v1
- Date: Tue, 11 May 2021 12:38:34 GMT
- Title: Tensor-Train Recurrent Neural Networks for Interpretable Multi-Way
Financial Forecasting
- Authors: Yao Lei Xu, Giuseppe G. Calvi, Danilo P. Mandic
- Abstract summary: Recurrent Neural Networks (RNNs) represent the de facto standard machine learning tool for sequence modelling.
The TT-RNN (TT-RNN) has the ability to deal with the curse of dimensionality, such as through the compression ability inherent to tensors.
We show, through the analysis of TT-factors, that the physical meaning underlying tensor decomposition, enables the TT-RNN model to aid the interpretability of results.
- Score: 24.50116388903113
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recurrent Neural Networks (RNNs) represent the de facto standard machine
learning tool for sequence modelling, owing to their expressive power and
memory. However, when dealing with large dimensional data, the corresponding
exponential increase in the number of parameters imposes a computational
bottleneck. The necessity to equip RNNs with the ability to deal with the curse
of dimensionality, such as through the parameter compression ability inherent
to tensors, has led to the development of the Tensor-Train RNN (TT-RNN).
Despite achieving promising results in many applications, the full potential of
the TT-RNN is yet to be explored in the context of interpretable financial
modelling, a notoriously challenging task characterized by multi-modal data
with low signal-to-noise ratio. To address this issue, we investigate the
potential of TT-RNN in the task of financial forecasting of currencies. We
show, through the analysis of TT-factors, that the physical meaning underlying
tensor decomposition, enables the TT-RNN model to aid the interpretability of
results, thus mitigating the notorious "black-box" issue associated with neural
networks. Furthermore, simulation results highlight the regularization power of
TT decomposition, demonstrating the superior performance of TT-RNN over its
uncompressed RNN counterpart and other tensor forecasting methods.
Related papers
- Scalable Mechanistic Neural Networks [52.28945097811129]
We propose an enhanced neural network framework designed for scientific machine learning applications involving long temporal sequences.
By reformulating the original Mechanistic Neural Network (MNN) we reduce the computational time and space complexities from cubic and quadratic with respect to the sequence length, respectively, to linear.
Extensive experiments demonstrate that S-MNN matches the original MNN in precision while substantially reducing computational resources.
arXiv Detail & Related papers (2024-10-08T14:27:28Z) - Is Conventional SNN Really Efficient? A Perspective from Network
Quantization [7.04833025737147]
Spiking Neural Networks (SNNs) have been widely praised for their high energy efficiency and immense potential.
However, comprehensive research that critically contrasts and correlates SNNs with quantized Artificial Neural Networks (ANNs) remains scant.
This paper introduces a unified perspective, illustrating that the time steps in SNNs and quantized bit-widths of activation values present analogous representations.
arXiv Detail & Related papers (2023-11-17T09:48:22Z) - Comparative Analysis of Interval Reachability for Robust Implicit and
Feedforward Neural Networks [64.23331120621118]
We use interval reachability analysis to obtain robustness guarantees for implicit neural networks (INNs)
INNs are a class of implicit learning models that use implicit equations as layers.
We show that our approach performs at least as well as, and generally better than, applying state-of-the-art interval bound propagation methods to INNs.
arXiv Detail & Related papers (2022-04-01T03:31:27Z) - Exploiting Low-Rank Tensor-Train Deep Neural Networks Based on
Riemannian Gradient Descent With Illustrations of Speech Processing [74.31472195046099]
We exploit a low-rank tensor-train deep neural network (TT-DNN) to build an end-to-end deep learning pipeline, namely LR-TT-DNN.
A hybrid model combining LR-TT-DNN with a convolutional neural network (CNN) is set up to boost the performance.
Our empirical evidence demonstrates that the LR-TT-DNN and CNN+(LR-TT-DNN) models with fewer model parameters can outperform the TT-DNN and CNN+(LR-TT-DNN) counterparts.
arXiv Detail & Related papers (2022-03-11T15:55:34Z) - On Recurrent Neural Networks for learning-based control: recent results
and ideas for future developments [1.1031750359996124]
This paper aims to discuss and analyze the potentialities of Recurrent Neural Networks (RNN) in control design.
Main families of RNN are considered, namely Neural AutoRegressive eXo, NNARX, Echo State Networks (ESN), Long Short Term Memory (LSTM), and Gated Recurrent Units (GRU)
arXiv Detail & Related papers (2021-11-26T15:52:52Z) - A Biased Graph Neural Network Sampler with Near-Optimal Regret [57.70126763759996]
Graph neural networks (GNN) have emerged as a vehicle for applying deep network architectures to graph and relational data.
In this paper, we build upon existing work and treat GNN neighbor sampling as a multi-armed bandit problem.
We introduce a newly-designed reward function that introduces some degree of bias designed to reduce variance and avoid unstable, possibly-unbounded payouts.
arXiv Detail & Related papers (2021-03-01T15:55:58Z) - Block-term Tensor Neural Networks [29.442026567710435]
We show that block-term tensor layers (BT-layers) can be easily adapted to neural network models, such as CNNs and RNNs.
BT-layers in CNNs and RNNs can achieve a very large compression ratio on the number of parameters while preserving or improving the representation power of the original DNNs.
arXiv Detail & Related papers (2020-10-10T09:58:43Z) - A Fully Tensorized Recurrent Neural Network [48.50376453324581]
We introduce a "fully tensorized" RNN architecture which jointly encodes the separate weight matrices within each recurrent cell.
This approach reduces model size by several orders of magnitude, while still maintaining similar or better performance compared to standard RNNs.
arXiv Detail & Related papers (2020-10-08T18:24:12Z) - Recurrent Graph Tensor Networks: A Low-Complexity Framework for
Modelling High-Dimensional Multi-Way Sequence [24.594587557319837]
We develop a graph filter framework for approximating the modelling of hidden states in Recurrent Neural Networks (RNNs)
The proposed framework is validated through several multi-way sequence modelling tasks and benchmarked against traditional RNNs.
We show that the proposed RGTN is capable of not only out-performing standard RNNs, but also mitigating the Curse of Dimensionality associated with traditional RNNs.
arXiv Detail & Related papers (2020-09-18T10:13:36Z) - Tensor train decompositions on recurrent networks [60.334946204107446]
Matrix product state (MPS) tensor trains have more attractive features than MPOs, in terms of storage reduction and computing time at inference.
We show that MPS tensor trains should be at the forefront of LSTM network compression through a theoretical analysis and practical experiments on NLP task.
arXiv Detail & Related papers (2020-06-09T18:25:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.