CLVSA: A Convolutional LSTM Based Variational Sequence-to-Sequence Model
with Attention for Predicting Trends of Financial Markets
- URL: http://arxiv.org/abs/2104.04041v1
- Date: Thu, 8 Apr 2021 20:31:04 GMT
- Title: CLVSA: A Convolutional LSTM Based Variational Sequence-to-Sequence Model
with Attention for Predicting Trends of Financial Markets
- Authors: Jia Wang, Tong Sun, Benyuan Liu, Yu Cao, Hongwei Zhu
- Abstract summary: We propose CLVSA, a hybrid model that captures variationally underlying features in raw financial trading data.
Our model outperforms basic models, such as convolutional neural network, vanilla LSTM network, and sequence-to-sequence model with attention.
Our experimental results show that, by introducing an approximate posterior, CLVSA takes advantage of an extra regularizer based on the Kullback-Leibler divergence to prevent itself from overfitting traps.
- Score: 12.020797636494267
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Financial markets are a complex dynamical system. The complexity comes from
the interaction between a market and its participants, in other words, the
integrated outcome of activities of the entire participants determines the
markets trend, while the markets trend affects activities of participants.
These interwoven interactions make financial markets keep evolving. Inspired by
stochastic recurrent models that successfully capture variability observed in
natural sequential data such as speech and video, we propose CLVSA, a hybrid
model that consists of stochastic recurrent networks, the sequence-to-sequence
architecture, the self- and inter-attention mechanism, and convolutional LSTM
units to capture variationally underlying features in raw financial trading
data. Our model outperforms basic models, such as convolutional neural network,
vanilla LSTM network, and sequence-to-sequence model with attention, based on
backtesting results of six futures from January 2010 to December 2017. Our
experimental results show that, by introducing an approximate posterior, CLVSA
takes advantage of an extra regularizer based on the Kullback-Leibler
divergence to prevent itself from overfitting traps.
Related papers
- MITA: Bridging the Gap between Model and Data for Test-time Adaptation [68.62509948690698]
Test-Time Adaptation (TTA) has emerged as a promising paradigm for enhancing the generalizability of models.
We propose Meet-In-The-Middle based MITA, which introduces energy-based optimization to encourage mutual adaptation of the model and data from opposing directions.
arXiv Detail & Related papers (2024-10-12T07:02:33Z) - Semi-Supervised Reward Modeling via Iterative Self-Training [52.48668920483908]
We propose Semi-Supervised Reward Modeling (SSRM), an approach that enhances RM training using unlabeled data.
We demonstrate that SSRM significantly improves reward models without incurring additional labeling costs.
Overall, SSRM substantially reduces the dependency on large volumes of human-annotated data, thereby decreasing the overall cost and time involved in training effective reward models.
arXiv Detail & Related papers (2024-09-10T22:57:58Z) - Copula Variational LSTM for High-dimensional Cross-market Multivariate
Dependence Modeling [46.75628526959982]
We make the first attempt to integrate variational sequential neural learning with copula-based dependence modeling.
Our variational neural network WPVC-VLSTM models variational sequential dependence degrees and structures across time series.
It outperforms benchmarks including linear models, volatility models, deep neural networks, and variational recurrent networks in cross-market portfolio forecasting.
arXiv Detail & Related papers (2023-05-09T08:19:08Z) - Safety-compliant Generative Adversarial Networks for Human Trajectory
Forecasting [95.82600221180415]
Human forecasting in crowds presents the challenges of modelling social interactions and outputting collision-free multimodal distribution.
We introduce SGANv2, an improved safety-compliant SGAN architecture equipped with motion-temporal interaction modelling and a transformer-based discriminator design.
arXiv Detail & Related papers (2022-09-25T15:18:56Z) - Bayesian Bilinear Neural Network for Predicting the Mid-price Dynamics
in Limit-Order Book Markets [84.90242084523565]
Traditional time-series econometric methods often appear incapable of capturing the true complexity of the multi-level interactions driving the price dynamics.
By adopting a state-of-the-art second-order optimization algorithm, we train a Bayesian bilinear neural network with temporal attention.
By addressing the use of predictive distributions to analyze errors and uncertainties associated with the estimated parameters and model forecasts, we thoroughly compare our Bayesian model with traditional ML alternatives.
arXiv Detail & Related papers (2022-03-07T18:59:54Z) - Dual-CLVSA: a Novel Deep Learning Approach to Predict Financial Markets
with Sentiment Measurements [11.97251638872227]
We propose a novel deep learning approach, named dual-CLVSA, to predict financial market movement with both trading data and the corresponding social sentiment measurements, each through a separate sequence-to-sequence channel.
The experiment results show that dual-CLVSA can effectively fuse the two types of data, and verify that sentiment measurements are not only informative for financial market predictions, but they also contain extra profitable features to boost the performance of our predicting system.
arXiv Detail & Related papers (2022-01-27T20:32:46Z) - Parsimonious Quantile Regression of Financial Asset Tail Dynamics via
Sequential Learning [35.34574502348672]
We propose a parsimonious quantile regression framework to learn the dynamic tail behaviors of financial asset returns.
Our model captures well both the time-varying characteristic and the asymmetrical heavy-tail property of financial time series.
arXiv Detail & Related papers (2020-10-16T09:35:52Z) - Markovian RNN: An Adaptive Time Series Prediction Network with HMM-based
Switching for Nonstationary Environments [11.716677452529114]
We introduce a novel recurrent neural network (RNN) architecture, which adaptively switches between internal regimes in a Markovian way to model the nonstationary nature of the given data.
Our model, Markovian RNN employs a hidden Markov model (HMM) for regime transitions, where each regime controls hidden state transitions of the recurrent cell independently.
We demonstrate the significant performance gains compared to vanilla RNN and conventional methods such as Markov Switching ARIMA.
arXiv Detail & Related papers (2020-06-17T19:38:29Z) - Gaussian process imputation of multiple financial series [71.08576457371433]
Multiple time series such as financial indicators, stock prices and exchange rates are strongly coupled due to their dependence on the latent state of the market.
We focus on learning the relationships among financial time series by modelling them through a multi-output Gaussian process.
arXiv Detail & Related papers (2020-02-11T19:18:18Z) - A Bayesian Long Short-Term Memory Model for Value at Risk and Expected
Shortfall Joint Forecasting [26.834110647177965]
Value-at-Risk (VaR) and Expected Shortfall (ES) are widely used in the financial sector to measure the market risk and manage the extreme market movement.
Recent link between the quantile score function and the Asymmetric Laplace density has led to a flexible likelihood-based framework for joint modelling of VaR and ES.
We develop a hybrid model that is based on the Asymmetric Laplace quasi-likelihood and employs the Long Short-Term Memory (LSTM) time series modelling technique from Machine Learning to capture efficiently the underlying dynamics of VaR and ES.
arXiv Detail & Related papers (2020-01-23T05:13:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.