Multi-Transformer: A New Neural Network-Based Architecture for
Forecasting S&P Volatility
- URL: http://arxiv.org/abs/2109.12621v1
- Date: Sun, 26 Sep 2021 14:47:04 GMT
- Title: Multi-Transformer: A New Neural Network-Based Architecture for
Forecasting S&P Volatility
- Authors: Eduardo Ramos-P\'erez, Pablo J. Alonso-Gonz\'alez, Jos\'e Javier
N\'u\~nez-Vel\'azquez
- Abstract summary: This paper proposes more accurate stock volatility models based on machine and deep learning techniques.
This paper introduces a neural network-based architecture, called Multi-Transformer.
The paper also adapts traditional Transformer layers in order to be used in volatility forecasting models.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Events such as the Financial Crisis of 2007-2008 or the COVID-19 pandemic
caused significant losses to banks and insurance entities. They also
demonstrated the importance of using accurate equity risk models and having a
risk management function able to implement effective hedging strategies. Stock
volatility forecasts play a key role in the estimation of equity risk and,
thus, in the management actions carried out by financial institutions.
Therefore, this paper has the aim of proposing more accurate stock volatility
models based on novel machine and deep learning techniques. This paper
introduces a neural network-based architecture, called Multi-Transformer.
Multi-Transformer is a variant of Transformer models, which have already been
successfully applied in the field of natural language processing. Indeed, this
paper also adapts traditional Transformer layers in order to be used in
volatility forecasting models. The empirical results obtained in this paper
suggest that the hybrid models based on Multi-Transformer and Transformer
layers are more accurate and, hence, they lead to more appropriate risk
measures than other autoregressive algorithms or hybrid models based on feed
forward layers or long short term memory cells.
Related papers
- Enhancing Risk Assessment in Transformers with Loss-at-Risk Functions [3.2162648244439684]
We introduce a novel loss function, the Loss-at-Risk, which incorporates Value at Risk (VaR) and Conditional Value at Risk (CVaR) into Transformer models.
This integration allows Transformer models to recognize potential extreme losses and further improves their capability to handle high-stakes financial decisions.
We conduct a series of experiments with highly volatile financial datasets to demonstrate that our Loss-at-Risk function improves the Transformers' risk prediction and management capabilities.
arXiv Detail & Related papers (2024-11-04T19:44:43Z) - Efficient Adaptation of Pre-trained Vision Transformer via Householder Transformation [53.88562288388169]
A common strategy for.
Efficient Fine-Tuning (PEFT) of pre-trained Vision Transformers (ViTs) involves adapting the model to downstream tasks.
We propose a novel PEFT approach inspired by Singular Value Decomposition (SVD) for representing the adaptation matrix.
SVD decomposes a matrix into the product of a left unitary matrix, a diagonal matrix of scaling values, and a right unitary matrix.
arXiv Detail & Related papers (2024-10-30T12:08:30Z) - Deep Analysis of Time Series Data for Smart Grid Startup Strategies: A Transformer-LSTM-PSO Model Approach [0.8702432681310401]
Transformer-LSTM-PSO model is designed to more effectively capture the complex temporal relationships in grid startup schemes.
Model achieves lower RMSE and MAE values across multiple datasets compared to existing benchmarks.
The application of the Transformer-LSTM-PSO model represents a significant advancement in smart grid predictive analytics.
arXiv Detail & Related papers (2024-08-22T04:52:02Z) - Enhancing Actuarial Non-Life Pricing Models via Transformers [0.0]
We build on the foundation laid out by the combined actuarial neural network as well as the localGLMnet and enhance those models via the feature tokenizer transformer.
The paper shows that the new methods can achieve better results than the benchmark models while preserving certain generalized linear model advantages.
arXiv Detail & Related papers (2023-11-10T12:06:23Z) - Differential Evolution Algorithm based Hyper-Parameters Selection of
Transformer Neural Network Model for Load Forecasting [0.0]
Transformer models have the potential to improve Load forecasting because of their ability to learn long-range dependencies derived from their Attention Mechanism.
Our work compares the proposed Transformer based Neural Network model integrated with different metaheuristic algorithms by their performance in Load forecasting based on numerical metrics such as Mean Squared Error (MSE) and Mean Absolute Percentage Error (MAPE)
arXiv Detail & Related papers (2023-07-28T04:29:53Z) - Emergent Agentic Transformer from Chain of Hindsight Experience [96.56164427726203]
We show that a simple transformer-based model performs competitively with both temporal-difference and imitation-learning-based approaches.
This is the first time that a simple transformer-based model performs competitively with both temporal-difference and imitation-learning-based approaches.
arXiv Detail & Related papers (2023-05-26T00:43:02Z) - Learning to Grow Pretrained Models for Efficient Transformer Training [72.20676008625641]
We learn to grow pretrained transformers, where we learn to linearly map the parameters of the smaller model to initialize the larger model.
Experiments across both language and vision transformers demonstrate that our learned Linear Growth Operator (LiGO) can save up to 50% computational cost of training from scratch.
arXiv Detail & Related papers (2023-03-02T05:21:18Z) - Full Stack Optimization of Transformer Inference: a Survey [58.55475772110702]
Transformer models achieve superior accuracy across a wide range of applications.
The amount of compute and bandwidth required for inference of recent Transformer models is growing at a significant rate.
There has been an increased focus on making Transformer models more efficient.
arXiv Detail & Related papers (2023-02-27T18:18:13Z) - Forecasting High-Dimensional Covariance Matrices of Asset Returns with
Hybrid GARCH-LSTMs [0.0]
This paper investigates the ability of hybrid models, mixing GARCH processes and neural networks, to forecast covariance matrices of asset returns.
The new model proposed is very promising as it not only outperforms the equally weighted portfolio, but also by a significant margin its econometric counterpart.
arXiv Detail & Related papers (2021-08-25T23:41:43Z) - Decision Transformer: Reinforcement Learning via Sequence Modeling [102.86873656751489]
We present a framework that abstracts Reinforcement Learning (RL) as a sequence modeling problem.
We present Decision Transformer, an architecture that casts the problem of RL as conditional sequence modeling.
Despite its simplicity, Decision Transformer matches or exceeds the performance of state-of-the-art offline RL baselines on Atari, OpenAI Gym, and Key-to-Door tasks.
arXiv Detail & Related papers (2021-06-02T17:53:39Z) - Bayesian Transformer Language Models for Speech Recognition [59.235405107295655]
State-of-the-art neural language models (LMs) represented by Transformers are highly complex.
This paper proposes a full Bayesian learning framework for Transformer LM estimation.
arXiv Detail & Related papers (2021-02-09T10:55:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.