Long short-term memory networks and laglasso for bond yield forecasting:
Peeping inside the black box
- URL: http://arxiv.org/abs/2005.02217v1
- Date: Tue, 5 May 2020 14:23:00 GMT
- Title: Long short-term memory networks and laglasso for bond yield forecasting:
Peeping inside the black box
- Authors: Manuel Nunes, Enrico Gerding, Frank McGroarty, Mahesan Niranjan
- Abstract summary: We conduct the first study of bond yield forecasting using long short-term memory (LSTM) networks.
We calculate the LSTM signals through time, at selected locations in the memory cell, using sequence-to-sequence architectures.
- Score: 10.412912723760172
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Modern decision-making in fixed income asset management benefits from
intelligent systems, which involve the use of state-of-the-art machine learning
models and appropriate methodologies. We conduct the first study of bond yield
forecasting using long short-term memory (LSTM) networks, validating its
potential and identifying its memory advantage. Specifically, we model the
10-year bond yield using univariate LSTMs with three input sequences and five
forecasting horizons. We compare those with multilayer perceptrons (MLP),
univariate and with the most relevant features. To demystify the notion of
black box associated with LSTMs, we conduct the first internal study of the
model. To this end, we calculate the LSTM signals through time, at selected
locations in the memory cell, using sequence-to-sequence architectures, uni and
multivariate. We then proceed to explain the states' signals using exogenous
information, for what we develop the LSTM-LagLasso methodology. The results
show that the univariate LSTM model with additional memory is capable of
achieving similar results as the multivariate MLP using macroeconomic and
market information. Furthermore, shorter forecasting horizons require smaller
input sequences and vice-versa. The most remarkable property found consistently
in the LSTM signals, is the activation/deactivation of units through time, and
the specialisation of units by yield range or feature. Those signals are
complex but can be explained by exogenous variables. Additionally, some of the
relevant features identified via LSTM-LagLasso are not commonly used in
forecasting models. In conclusion, our work validates the potential of LSTMs
and methodologies for bonds, providing additional tools for financial
practitioners.
Related papers
- Unlocking the Power of LSTM for Long Term Time Series Forecasting [27.245021350821638]
We propose a simple yet efficient algorithm named P-sLSTM built upon sLSTM by incorporating patching and channel independence.
These modifications substantially enhance sLSTM's performance in TSF, achieving state-of-the-art results.
arXiv Detail & Related papers (2024-08-19T13:59:26Z) - Beam Prediction based on Large Language Models [51.45077318268427]
Millimeter-wave (mmWave) communication is promising for next-generation wireless networks but suffers from significant path loss.
Traditional deep learning models, such as long short-term memory (LSTM), enhance beam tracking accuracy however are limited by poor robustness and generalization.
In this letter, we use large language models (LLMs) to improve the robustness of beam prediction.
arXiv Detail & Related papers (2024-08-16T12:40:01Z) - HOPE for a Robust Parameterization of Long-memory State Space Models [51.66430224089725]
State-space models (SSMs) that utilize linear, time-invariant (LTI) systems are known for their effectiveness in learning long sequences.
We develop a new parameterization scheme, called HOPE, for LTI systems that utilize Markov parameters within Hankel operators.
Our new parameterization endows the SSM with non-decaying memory within a fixed time window, which is empirically corroborated by a sequential CIFAR-10 task with padded noise.
arXiv Detail & Related papers (2024-05-22T20:20:14Z) - Federated Quantum Long Short-term Memory (FedQLSTM) [58.50321380769256]
Quantum federated learning (QFL) can facilitate collaborative learning across multiple clients using quantum machine learning (QML) models.
No prior work has focused on developing a QFL framework that utilizes temporal data to approximate functions.
A novel QFL framework that is the first to integrate quantum long short-term memory (QLSTM) models with temporal data is proposed.
arXiv Detail & Related papers (2023-12-21T21:40:47Z) - DeLELSTM: Decomposition-based Linear Explainable LSTM to Capture
Instantaneous and Long-term Effects in Time Series [26.378073712630467]
We propose a Decomposition-based Linear Explainable LSTM (DeLELSTM) to improve the interpretability of LSTM.
We demonstrate the effectiveness and interpretability of DeLELSTM on three empirical datasets.
arXiv Detail & Related papers (2023-08-26T07:45:41Z) - Hybrid hidden Markov LSTM for short-term traffic flow prediction [0.0]
We propose a hybrid hidden Markov-LSTM model that is capable of learning complementary features in traffic data.
Results indicate significant performance gains in using hybrid architecture compared to conventional methods.
arXiv Detail & Related papers (2023-07-11T00:56:44Z) - Working Memory Connections for LSTM [51.742526187978726]
We show that Working Memory Connections constantly improve the performance of LSTMs on a variety of tasks.
Numerical results suggest that the cell state contains useful information that is worth including in the gate structure.
arXiv Detail & Related papers (2021-08-31T18:01:30Z) - Learning representations with end-to-end models for improved remaining
useful life prognostics [64.80885001058572]
The remaining Useful Life (RUL) of equipment is defined as the duration between the current time and its failure.
We propose an end-to-end deep learning model based on multi-layer perceptron and long short-term memory layers (LSTM) to predict the RUL.
We will discuss how the proposed end-to-end model is able to achieve such good results and compare it to other deep learning and state-of-the-art methods.
arXiv Detail & Related papers (2021-04-11T16:45:18Z) - Quantum Long Short-Term Memory [3.675884635364471]
Long short-term memory (LSTM) is a recurrent neural network (RNN) for sequence and temporal dependency data modeling.
We propose a hybrid quantum-classical model of LSTM, which we dub QLSTM.
Our work paves the way toward implementing machine learning algorithms for sequence modeling on noisy intermediate-scale quantum (NISQ) devices.
arXiv Detail & Related papers (2020-09-03T16:41:09Z) - Automatic Remaining Useful Life Estimation Framework with Embedded
Convolutional LSTM as the Backbone [5.927250637620123]
We propose a new LSTM variant called embedded convolutional LSTM (E NeuralTM)
In ETM a group of different 1D convolutions is embedded into the LSTM structure. Through this, the temporal information is preserved between and within windows.
We show the superiority of our proposed ETM approach over the state-of-the-art approaches on several widely used benchmark data sets for RUL Estimation.
arXiv Detail & Related papers (2020-08-10T08:34:20Z) - Deep Stock Predictions [58.720142291102135]
We consider the design of a trading strategy that performs portfolio optimization using Long Short Term Memory (LSTM) neural networks.
We then customize the loss function used to train the LSTM to increase the profit earned.
We find the LSTM model with the customized loss function to have an improved performance in the training bot over a regressive baseline such as ARIMA.
arXiv Detail & Related papers (2020-06-08T23:37:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.