Attention-based CNN-LSTM and XGBoost hybrid model for stock prediction
- URL: http://arxiv.org/abs/2204.02623v1
- Date: Wed, 6 Apr 2022 07:06:30 GMT
- Title: Attention-based CNN-LSTM and XGBoost hybrid model for stock prediction
- Authors: Zhuangwei Shi, Yang Hu, Guangliang Mo, Jian Wu
- Abstract summary: This paper proposes an attention-based CNN-LSTM and XGBoost hybrid model to predict the stock price.
The model can fully mine the historical information of the stock market in multiple periods.
The results show that the hybrid model is more effective and the prediction accuracy is relatively high.
- Score: 7.231134145443057
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Stock market plays an important role in the economic development. Due to the
complex volatility of the stock market, the research and prediction on the
change of the stock price, can avoid the risk for the investors. The
traditional time series model ARIMA can not describe the nonlinearity, and can
not achieve satisfactory results in the stock prediction. As neural networks
are with strong nonlinear generalization ability, this paper proposes an
attention-based CNN-LSTM and XGBoost hybrid model to predict the stock price.
The model constructed in this paper integrates the time series model, the
Convolutional Neural Networks with Attention mechanism, the Long Short-Term
Memory network, and XGBoost regressor in a non-linear relationship, and
improves the prediction accuracy. The model can fully mine the historical
information of the stock market in multiple periods. The stock data is first
preprocessed through ARIMA. Then, the deep learning architecture formed in
pretraining-finetuning framework is adopted. The pre-training model is the
Attention-based CNN-LSTM model based on sequence-to-sequence framework. The
model first uses convolution to extract the deep features of the original stock
data, and then uses the Long Short-Term Memory networks to mine the long-term
time series features. Finally, the XGBoost model is adopted for fine-tuning.
The results show that the hybrid model is more effective and the prediction
accuracy is relatively high, which can help investors or institutions to make
decisions and achieve the purpose of expanding return and avoiding risk. Source
code is available at
https://github.com/zshicode/Attention-CLX-stock-prediction.
Related papers
- GARCH-Informed Neural Networks for Volatility Prediction in Financial Markets [0.0]
We present a new, hybrid Deep Learning model that captures and forecasting market volatility more accurately than either class of models are capable of on their own.
When compared to other time series models, GINN showed superior out-of-sample prediction performance in terms of the Coefficient of Determination ($R2$), Mean Squared Error (MSE), and Mean Absolute Error (MAE)
arXiv Detail & Related papers (2024-09-30T23:53:54Z) - F-FOMAML: GNN-Enhanced Meta-Learning for Peak Period Demand Forecasting with Proxy Data [65.6499834212641]
We formulate the demand prediction as a meta-learning problem and develop the Feature-based First-Order Model-Agnostic Meta-Learning (F-FOMAML) algorithm.
By considering domain similarities through task-specific metadata, our model improved generalization, where the excess risk decreases as the number of training tasks increases.
Compared to existing state-of-the-art models, our method demonstrates a notable improvement in demand prediction accuracy, reducing the Mean Absolute Error by 26.24% on an internal vending machine dataset and by 1.04% on the publicly accessible JD.com dataset.
arXiv Detail & Related papers (2024-06-23T21:28:50Z) - Gated recurrent neural network with TPE Bayesian optimization for enhancing stock index prediction accuracy [0.0]
The aim is to improve the prediction accuracy of the next day's closing price of the NIFTY 50 index, a prominent Indian stock market index.
A combination of eight influential factors is carefully chosen from fundamental stock data, technical indicators, crude oil price, and macroeconomic data to train the models.
arXiv Detail & Related papers (2024-06-02T06:39:01Z) - A Study on Stock Forecasting Using Deep Learning and Statistical Models [3.437407981636465]
This paper will review many deep learning algorithms for stock price forecasting. We use a record of s&p 500 index data for training and testing.
It will discuss various models, including the Auto regression integration moving average model, the Recurrent neural network model, the long short-term model, the convolutional neural network model, and the full convolutional neural network model.
arXiv Detail & Related papers (2024-02-08T16:45:01Z) - A Dynamical Model of Neural Scaling Laws [79.59705237659547]
We analyze a random feature model trained with gradient descent as a solvable model of network training and generalization.
Our theory shows how the gap between training and test loss can gradually build up over time due to repeated reuse of data.
arXiv Detail & Related papers (2024-02-02T01:41:38Z) - Diffusion Variational Autoencoder for Tackling Stochasticity in
Multi-Step Regression Stock Price Prediction [54.21695754082441]
Multi-step stock price prediction over a long-term horizon is crucial for forecasting its volatility.
Current solutions to multi-step stock price prediction are mostly designed for single-step, classification-based predictions.
We combine a deep hierarchical variational-autoencoder (VAE) and diffusion probabilistic techniques to do seq2seq stock prediction.
Our model is shown to outperform state-of-the-art solutions in terms of its prediction accuracy and variance.
arXiv Detail & Related papers (2023-08-18T16:21:15Z) - nanoLM: an Affordable LLM Pre-training Benchmark via Accurate Loss Prediction across Scales [65.01417261415833]
We present an approach to predict the pre-training loss based on our observations that Maximal Update Parametrization (muP) enables accurate fitting of scaling laws.
With around 14% of the one-time pre-training cost, we can accurately forecast the loss for models up to 52B.
Our goal with nanoLM is to empower researchers with limited resources to reach meaningful conclusions on large models.
arXiv Detail & Related papers (2023-04-14T00:45:01Z) - Augmented Bilinear Network for Incremental Multi-Stock Time-Series
Classification [83.23129279407271]
We propose a method to efficiently retain the knowledge available in a neural network pre-trained on a set of securities.
In our method, the prior knowledge encoded in a pre-trained neural network is maintained by keeping existing connections fixed.
This knowledge is adjusted for the new securities by a set of augmented connections, which are optimized using the new data.
arXiv Detail & Related papers (2022-07-23T18:54:10Z) - Generative Adversarial Network (GAN) and Enhanced Root Mean Square Error
(ERMSE): Deep Learning for Stock Price Movement Prediction [15.165487282631535]
This paper aims to improve prediction accuracy and minimize forecasting error loss by using Generative Adversarial Networks.
It was found that the Generative Adversarial Network (GAN) has performed well on the enhanced root mean square error to LSTM.
arXiv Detail & Related papers (2021-11-30T18:38:59Z) - Stock Price Prediction Using CNN and LSTM-Based Deep Learning Models [0.0]
This paper presents a suite of deep learning based models for stock price prediction.
We use the historical records of the NIFTY 50 index listed in the National Stock Exchange of India.
Our proposition includes two regression models built on convolutional neural networks and three long and short term memory network based predictive models.
arXiv Detail & Related papers (2020-10-22T03:09:07Z) - Deep Stock Predictions [58.720142291102135]
We consider the design of a trading strategy that performs portfolio optimization using Long Short Term Memory (LSTM) neural networks.
We then customize the loss function used to train the LSTM to increase the profit earned.
We find the LSTM model with the customized loss function to have an improved performance in the training bot over a regressive baseline such as ARIMA.
arXiv Detail & Related papers (2020-06-08T23:37:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.