Interpretable Machine Learning for Macro Alpha: A News Sentiment Case Study
- URL: http://arxiv.org/abs/2505.16136v1
- Date: Thu, 22 May 2025 02:24:45 GMT
- Title: Interpretable Machine Learning for Macro Alpha: A News Sentiment Case Study
- Authors: Yuke Zhang,
- Abstract summary: We process the Global Database of Events, Language, and Tone (GDELT) Project's worldwide news feed using FinBERT.<n>We construct daily sentiment indices incorporating mean tone, dispersion, and event impact.<n>These indices drive an XGBoost, benchmarked against logistic regression, to predict next-day returns.
- Score: 1.57731592348751
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This study introduces an interpretable machine learning (ML) framework to extract macroeconomic alpha from global news sentiment. We process the Global Database of Events, Language, and Tone (GDELT) Project's worldwide news feed using FinBERT -- a Bidirectional Encoder Representations from Transformers (BERT) based model pretrained on finance-specific language -- to construct daily sentiment indices incorporating mean tone, dispersion, and event impact. These indices drive an XGBoost classifier, benchmarked against logistic regression, to predict next-day returns for EUR/USD, USD/JPY, and 10-year U.S. Treasury futures (ZN). Rigorous out-of-sample (OOS) backtesting (5-fold expanding-window cross-validation, OOS period: c. 2017-April 2025) demonstrates exceptional, cost-adjusted performance for the XGBoost strategy: Sharpe ratios achieve 5.87 (EUR/USD), 4.65 (USD/JPY), and 4.65 (Treasuries), with respective compound annual growth rates (CAGRs) exceeding 50% in Foreign Exchange (FX) and 22% in bonds. Shapley Additive Explanations (SHAP) affirm that sentiment dispersion and article impact are key predictive features. Our findings establish that integrating domain-specific Natural Language Processing (NLP) with interpretable ML offers a potent and explainable source of macro alpha.
Related papers
- PEGRL: Improving Machine Translation by Post-Editing Guided Reinforcement Learning [54.19784655270799]
We introduce textbfPEGRL, a textittwo-stage RL framework that uses post-editing as an auxiliary task to stabilize training and guide overall optimization.<n>Experiments on English$to$Finnish, English$to$Turkish, and English$leftrightarrow$Chinese show consistent gains over RL baselines.
arXiv Detail & Related papers (2026-02-03T10:22:55Z) - PriceSeer: Evaluating Large Language Models in Real-Time Stock Prediction [47.70107097572211]
We introduce PriceSeer, a benchmark specifically designed for large language models performing stock prediction tasks.<n>PriceSeer includes 110 U.S. stocks from 11 industrial sectors, with each containing 249 historical data points.<n>We evaluate six cutting-edge LLMs under different prediction horizons, demonstrating their potential in generating investment strategies.
arXiv Detail & Related papers (2025-12-31T08:35:46Z) - Can We Reliably Predict the Fed's Next Move? A Multi-Modal Approach to U.S. Monetary Policy Forecasting [2.6396287656676733]
This study examines whether predictive accuracy can be enhanced by integrating structured data with unstructured textual signals from Federal Reserve communications.<n>Our results show that hybrid models consistently outperform unimodal baselines.<n>For monetary policy forecasting, simpler hybrid models can offer both accuracy and interpretability, delivering actionable insights for researchers and decision-makers.
arXiv Detail & Related papers (2025-06-28T05:54:58Z) - Explainable-AI powered stock price prediction using time series transformers: A Case Study on BIST100 [0.0]
This study proposes a novel approach that combines transformer-based time series models with explainable artificial intelligence (XAI) to enhance the interpretability and accuracy of stock price predictions.<n>The analysis focuses on the daily stock prices of the five highest-volume banks listed in the BIST100 index, along with XBANK and XU100 indices, covering the period from January 2015 to March 2025.
arXiv Detail & Related papers (2025-06-01T13:29:25Z) - Innovative Sentiment Analysis and Prediction of Stock Price Using FinBERT, GPT-4 and Logistic Regression: A Data-Driven Approach [0.0]
This study explores the comparative performance of cutting-edge AI models, i.e., Finaance Bidirectional representations from Transsformers (FinBERT), Generatice Pre-trained Transformer GPT-4, and Logistic Regression, for sentiment analysis and stock index prediction.<n>By leveraging advanced natural language processing models like GPT-4 and FinBERT, alongside a traditional machine learning model, Logistic Regression, we aim to classify market sentiment, generate sentiment scores, and predict market price movements.
arXiv Detail & Related papers (2024-12-07T05:20:31Z) - BreakGPT: Leveraging Large Language Models for Predicting Asset Price Surges [55.2480439325792]
This paper introduces BreakGPT, a novel large language model (LLM) architecture adapted specifically for time series forecasting and the prediction of sharp upward movements in asset prices.
We showcase BreakGPT as a promising solution for financial forecasting with minimal training and as a strong competitor for capturing both local and global temporal dependencies.
arXiv Detail & Related papers (2024-11-09T05:40:32Z) - Harnessing Earnings Reports for Stock Predictions: A QLoRA-Enhanced LLM Approach [6.112119533910774]
This paper introduces an advanced approach by employing Large Language Models (LLMs) instruction fine-tuned with a novel combination of instruction-based techniques and quantized low-rank adaptation (QLoRA) compression.
Our methodology integrates 'base factors', such as financial metric growth and earnings transcripts, with 'external factors', including recent market indices performances and analyst grades, to create a rich, supervised dataset.
This study not only demonstrates the power of integrating cutting-edge AI with fine-tuned financial data but also paves the way for future research in enhancing AI-driven financial analysis tools.
arXiv Detail & Related papers (2024-08-13T04:53:31Z) - Advancing Anomaly Detection: Non-Semantic Financial Data Encoding with LLMs [49.57641083688934]
We introduce a novel approach to anomaly detection in financial data using Large Language Models (LLMs) embeddings.
Our experiments demonstrate that LLMs contribute valuable information to anomaly detection as our models outperform the baselines.
arXiv Detail & Related papers (2024-06-05T20:19:09Z) - CIF-Bench: A Chinese Instruction-Following Benchmark for Evaluating the Generalizability of Large Language Models [53.9835961434552]
We introduce the Chinese Instruction-Following Benchmark (CIF-Bench) to evaluate the generalizability of large language models (LLMs) to the Chinese language.
CIF-Bench comprises 150 tasks and 15,000 input-output pairs, developed by native speakers to test complex reasoning and Chinese cultural nuances.
To mitigate data contamination, we release only half of the dataset publicly, with the remainder kept private, and introduce diversified instructions to minimize score variance.
arXiv Detail & Related papers (2024-02-20T16:02:12Z) - Ploutos: Towards interpretable stock movement prediction with financial
large language model [43.51934592920784]
Ploutos is a novel financial framework that consists of PloutosGen and PloutosGPT.
The PloutosGen contains multiple primary experts that can analyze different modal data, such as text and numbers, and provide quantitative strategies from different perspectives.
The training strategy of PloutosGPT leverage rearview-mirror prompting mechanism to guide GPT-4 to generate rationales, and a dynamic token weighting mechanism to finetune LLM.
arXiv Detail & Related papers (2024-02-18T10:28:18Z) - Can Large Language Models Beat Wall Street? Unveiling the Potential of AI in Stock Selection [0.07499722271664146]
This paper introduces MarketSenseAI, an innovative framework leveraging GPT-4's advanced reasoning for selecting stocks in financial markets.
MarketSenseAI analyzes diverse data sources, including market trends, news, fundamentals, and macroeconomic factors, to emulate expert investment decision-making.
Our findings highlight the transformative potential of Large Language Models in financial decision-making.
arXiv Detail & Related papers (2024-01-08T08:58:46Z) - Extrapolating Large Language Models to Non-English by Aligning Languages [109.09051737966178]
Existing large language models show disparate capability across different languages.
In this paper, we empower pre-trained LLMs on non-English languages by building semantic alignment across languages.
arXiv Detail & Related papers (2023-08-09T13:32:06Z) - Can ChatGPT Forecast Stock Price Movements? Return Predictability and Large Language Models [51.3422222472898]
We document the capability of large language models (LLMs) like ChatGPT to predict stock price movements using news headlines.
We develop a theoretical model incorporating information capacity constraints, underreaction, limits-to-arbitrage, and LLMs.
arXiv Detail & Related papers (2023-04-15T19:22:37Z) - Accelerating Vision-Language Pretraining with Free Language Modeling [62.30042851111692]
Free language modeling (FLM) enables a 100% prediction rate with arbitrary corruption rates.
FLM successfully frees the prediction rate from the tie-up with the corruption rate.
Experiments show FLM could achieve an impressive 2.5x pretraining time reduction.
arXiv Detail & Related papers (2023-03-24T14:49:22Z) - BBT-Fin: Comprehensive Construction of Chinese Financial Domain
Pre-trained Language Model, Corpus and Benchmark [12.457193087920183]
We introduce BBT-FinT5, a new Chinese financial pre-training language model based on the T5 model.
To support this effort, we have built BBT-FinCorpus, a large-scale financial corpus with approximately 300GB of raw text from four different sources.
arXiv Detail & Related papers (2023-02-18T22:20:37Z) - Explicit Alignment Objectives for Multilingual Bidirectional Encoders [111.65322283420805]
We present a new method for learning multilingual encoders, AMBER (Aligned Multilingual Bi-directional EncodeR)
AMBER is trained on additional parallel data using two explicit alignment objectives that align the multilingual representations at different granularities.
Experimental results show that AMBER obtains gains of up to 1.1 average F1 score on sequence tagging and up to 27.3 average accuracy on retrieval over the XLMR-large model.
arXiv Detail & Related papers (2020-10-15T18:34:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.