Improved Stock Price Movement Classification Using News Articles Based
on Embeddings and Label Smoothing
- URL: http://arxiv.org/abs/2301.10458v1
- Date: Wed, 25 Jan 2023 08:33:45 GMT
- Title: Improved Stock Price Movement Classification Using News Articles Based
on Embeddings and Label Smoothing
- Authors: Luis Villamil, Ryan Bausback, Shaeke Salman, Ting L. Liu, Conrad Horn,
Xiuwen Liu
- Abstract summary: We propose to improve stock price movement classification using news articles by incorporating regularization and optimization techniques from deep learning.
We further incorporate weight decay, batch normalization, dropout, and label smoothing to improve the generalization of the trained models.
Our experimental results on a commonly used dataset show significant improvements, achieving average accuracy of 80.7% on the test set.
- Score: 1.8920934738244022
- License: http://creativecommons.org/publicdomain/zero/1.0/
- Abstract: Stock price movement prediction is a challenging and essential problem in
finance. While it is well established in modern behavioral finance that the
share prices of related stocks often move after the release of news via
reactions and overreactions of investors, how to capture the relationships
between price movements and news articles via quantitative models is an active
area research; existing models have achieved success with variable degrees. In
this paper, we propose to improve stock price movement classification using
news articles by incorporating regularization and optimization techniques from
deep learning. More specifically, we capture the dependencies between news
articles and stocks through embeddings and bidirectional recurrent neural
networks as in recent models. We further incorporate weight decay, batch
normalization, dropout, and label smoothing to improve the generalization of
the trained models. To handle high fluctuations of validation accuracy of batch
normalization, we propose dual-phase training to realize the improvements
reliably. Our experimental results on a commonly used dataset show significant
improvements, achieving average accuracy of 80.7% on the test set, which is
more than 10.0% absolute improvement over existing models. Our ablation studies
show batch normalization and label smoothing are most effective, leading to
6.0% and 3.4% absolute improvement, respectively on average.
Related papers
- TLOB: A Novel Transformer Model with Dual Attention for Stock Price Trend Prediction with Limit Order Book Data [10.684577067675585]
Stock Price Trend Prediction (SPTP) based on Limit Order Book (LOB) data is a fundamental challenge in financial markets.
Despite advances in deep learning, existing models fail to generalize across different market conditions.
We show that we surpass SoTA performance; thus, challenging the necessity of a complex spatial architecture.
arXiv Detail & Related papers (2025-02-12T12:41:10Z) - Re-evaluating Group Robustness via Adaptive Class-Specific Scaling [47.41034887474166]
Group distributionally robust optimization is a prominent algorithm used to mitigate spurious correlations and address dataset bias.
Existing approaches have reported improvements in robust accuracies but come at the cost of average accuracy due to inherent trade-offs.
We propose a class-specific scaling strategy, directly applicable to existing debiasing algorithms with no additional training.
We develop an instance-wise adaptive scaling technique to alleviate this trade-off, even leading to improvements in both robust and average accuracies.
arXiv Detail & Related papers (2024-12-19T16:01:51Z) - FinGPT: Enhancing Sentiment-Based Stock Movement Prediction with Dissemination-Aware and Context-Enriched LLMs [8.309532607378896]
We propose a data-driven approach that enhances sentiment-based stock movement predictions by incorporating news dissemination breadth, contextual data, and explicit instructions.
Our approach improves prediction accuracy by 8% compared to existing methods.
arXiv Detail & Related papers (2024-12-14T13:04:42Z) - AI in Investment Analysis: LLMs for Equity Stock Ratings [0.2916558661202724]
This paper explores the application of Large Language Models (LLMs) to generate multi-horizon stock ratings.
Our study addresses these issues by leveraging LLMs to improve the accuracy and consistency of stock ratings.
Our results show that our benchmark method outperforms traditional stock rating methods when assessed by forward returns.
arXiv Detail & Related papers (2024-10-30T15:06:57Z) - Optimizing Portfolio Management and Risk Assessment in Digital Assets
Using Deep Learning for Predictive Analysis [5.015409508372732]
This paper introduces the DQN algorithm into asset management portfolios in a novel and straightforward way.
The performance greatly exceeds the benchmark, which fully proves the effectiveness of the DRL algorithm in portfolio management.
Since different assets are trained separately as environments, there may be a phenomenon of Q value drift among different assets.
arXiv Detail & Related papers (2024-02-25T05:23:57Z) - BAL: Balancing Diversity and Novelty for Active Learning [53.289700543331925]
We introduce a novel framework, Balancing Active Learning (BAL), which constructs adaptive sub-pools to balance diverse and uncertain data.
Our approach outperforms all established active learning methods on widely recognized benchmarks by 1.20%.
arXiv Detail & Related papers (2023-12-26T08:14:46Z) - Towards systematic intraday news screening: a liquidity-focused approach [1.688090639493357]
Given the huge amount of news articles published each day, most of which are neutral, we present a systematic news screening method to identify the true'' impactful ones.
We show that the screened dataset leads to more effective feature capturing and thus superior performance on short-term asset return prediction.
arXiv Detail & Related papers (2023-04-11T10:14:48Z) - TWINS: A Fine-Tuning Framework for Improved Transferability of
Adversarial Robustness and Generalization [89.54947228958494]
This paper focuses on the fine-tuning of an adversarially pre-trained model in various classification tasks.
We propose a novel statistics-based approach, Two-WIng NormliSation (TWINS) fine-tuning framework.
TWINS is shown to be effective on a wide range of image classification datasets in terms of both generalization and robustness.
arXiv Detail & Related papers (2023-03-20T14:12:55Z) - Augmented Bilinear Network for Incremental Multi-Stock Time-Series
Classification [83.23129279407271]
We propose a method to efficiently retain the knowledge available in a neural network pre-trained on a set of securities.
In our method, the prior knowledge encoded in a pre-trained neural network is maintained by keeping existing connections fixed.
This knowledge is adjusted for the new securities by a set of augmented connections, which are optimized using the new data.
arXiv Detail & Related papers (2022-07-23T18:54:10Z) - Towards Reducing Labeling Cost in Deep Object Detection [61.010693873330446]
We propose a unified framework for active learning, that considers both the uncertainty and the robustness of the detector.
Our method is able to pseudo-label the very confident predictions, suppressing a potential distribution drift.
arXiv Detail & Related papers (2021-06-22T16:53:09Z) - A Sentiment Analysis Approach to the Prediction of Market Volatility [62.997667081978825]
We have explored the relationship between sentiment extracted from financial news and tweets and FTSE100 movements.
The sentiment captured from news headlines could be used as a signal to predict market returns; the same does not apply for volatility.
We developed an accurate classifier for the prediction of market volatility in response to the arrival of new information.
arXiv Detail & Related papers (2020-12-10T01:15:48Z) - Evaluating Prediction-Time Batch Normalization for Robustness under
Covariate Shift [81.74795324629712]
We call prediction-time batch normalization, which significantly improves model accuracy and calibration under covariate shift.
We show that prediction-time batch normalization provides complementary benefits to existing state-of-the-art approaches for improving robustness.
The method has mixed results when used alongside pre-training, and does not seem to perform as well under more natural types of dataset shift.
arXiv Detail & Related papers (2020-06-19T05:08:43Z) - Deep Learning for Portfolio Optimization [5.833272638548154]
Instead of selecting individual assets, we trade Exchange-Traded Funds (ETFs) of market indices to form a portfolio.
We compare our method with a wide range of algorithms with results showing that our model obtains the best performance over the testing period.
arXiv Detail & Related papers (2020-05-27T21:28:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.