Arbitrage-Free Bond and Yield Curve Forecasting with Neural Filters under HJM Constraints
- URL: http://arxiv.org/abs/2511.17892v1
- Date: Sat, 22 Nov 2025 02:47:27 GMT
- Title: Arbitrage-Free Bond and Yield Curve Forecasting with Neural Filters under HJM Constraints
- Authors: Xiang Gao, Cody Hyndman,
- Abstract summary: We develop an arbitrage-free deep learning framework for yield curve and bond price forecasting based on the Heath-Jarrow-Morton (HJM) term-structure model.<n>Our approach embeds a no-arbitrage drift restriction into a neural state-space architecture by combining Kalman, extended Kalman, and particle filters with recurrent neural networks (LSTM/TM)
- Score: 4.311211660681507
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We develop an arbitrage-free deep learning framework for yield curve and bond price forecasting based on the Heath-Jarrow-Morton (HJM) term-structure model and a dynamic Nelson-Siegel parameterization of forward rates. Our approach embeds a no-arbitrage drift restriction into a neural state-space architecture by combining Kalman, extended Kalman, and particle filters with recurrent neural networks (LSTM/CLSTM), and introduces an explicit arbitrage error regularization (AER) term during training. The model is applied to U.S. Treasury and corporate bond data, and its performance is evaluated for both yield-space and price-space predictions at 1-day and 5-day horizons. Empirically, arbitrage regularization leads to its strongest improvements at short maturities, particularly in 5-day-ahead forecasts, increasing market-consistency as measured by bid-ask hit rates and reducing dollar-denominated prediction errors.
Related papers
- Exploring the Interpretability of Forecasting Models for Energy Balancing Market [43.548887305614585]
The balancing market in the energy sector plays a critical role in physically and financially balancing the supply and demand.<n>While complex machine learning models can achieve high accuracy, their black-box nature severely limits the model interpretability.<n>This paper explores the trade-off between model accuracy and interpretability for the energy balancing market.
arXiv Detail & Related papers (2026-01-19T12:56:41Z) - Robust Yield Curve Estimation for Mortgage Bonds Using Neural Networks [0.6437467451172677]
We propose a neural networkbased framework for robust yield curve estimation tailored to small mortgage bond markets.<n>Our model estimates the yield curve independently for each day and introduces a new loss function to enforce smoothness and stability.<n> Empirical results on Swedish mortgage bonds demonstrate that our approach delivers more robust and stable yield curve estimates.
arXiv Detail & Related papers (2025-10-24T11:24:41Z) - Neural ARFIMA model for forecasting BRIC exchange rates with long memory under oil shocks and policy uncertainties [0.0]
Key drivers of exchange rate dynamics include global economic policy uncertainty, US equity market volatility, US monetary policy uncertainty, oil price growth rates, and country-specific short-term interest rate differentials.<n>We propose a Neural AutoRegressive Fractionally Integrated Moving Average model that combines the long-memory representation of ARFIMA with the nonlinear learning capacity of neural networks.<n>We show that NARFIMA consistently outperforms various state-of-the-art statistical and machine learning models in forecasting BRIC exchange rates.
arXiv Detail & Related papers (2025-09-08T13:49:48Z) - PREIG: Physics-informed and Reinforcement-driven Interpretable GRU for Commodity Demand Forecasting [27.542312745632458]
PREIG is a novel deep learning framework tailored for commodity demand forecasting.<n>The model uniquely integrates a Gated Recurrent Unit (GRU) architecture with physics-informed neural network (PINN) principles.
arXiv Detail & Related papers (2025-07-29T11:38:07Z) - GARCH-Informed Neural Networks for Volatility Prediction in Financial Markets [0.0]
We present a new, hybrid Deep Learning model that captures and forecasting market volatility more accurately than either class of models are capable of on their own.
When compared to other time series models, GINN showed superior out-of-sample prediction performance in terms of the Coefficient of Determination ($R2$), Mean Squared Error (MSE), and Mean Absolute Error (MAE)
arXiv Detail & Related papers (2024-09-30T23:53:54Z) - Inside the black box: Neural network-based real-time prediction of US recessions [0.0]
Long short-term memory (LSTM) and gated recurrent unit (GRU) are used to model US recessions from 1967 to 2021.
Shap method delivers key recession indicators, such as the S&P 500 index for short-term forecasting up to 3 months.
arXiv Detail & Related papers (2023-10-26T16:58:16Z) - Combining Deep Learning and GARCH Models for Financial Volatility and
Risk Forecasting [0.0]
We develop a hybrid approach to forecasting the volatility and risk of financial instruments by combining common econometric GARCH time series models with deep learning neural networks.
For the latter, we employ Gated Recurrent Unit (GRU) networks, whereas four different specifications are used as the GARCH component: standard GARCH, EGARCH, GJR-GARCH and APARCH.
Models are tested using daily logarithmic returns on the S&P 500 index as well as gold price Bitcoin prices.
arXiv Detail & Related papers (2023-10-02T10:18:13Z) - Diffusion Variational Autoencoder for Tackling Stochasticity in
Multi-Step Regression Stock Price Prediction [54.21695754082441]
Multi-step stock price prediction over a long-term horizon is crucial for forecasting its volatility.
Current solutions to multi-step stock price prediction are mostly designed for single-step, classification-based predictions.
We combine a deep hierarchical variational-autoencoder (VAE) and diffusion probabilistic techniques to do seq2seq stock prediction.
Our model is shown to outperform state-of-the-art solutions in terms of its prediction accuracy and variance.
arXiv Detail & Related papers (2023-08-18T16:21:15Z) - Bayesian Bilinear Neural Network for Predicting the Mid-price Dynamics
in Limit-Order Book Markets [84.90242084523565]
Traditional time-series econometric methods often appear incapable of capturing the true complexity of the multi-level interactions driving the price dynamics.
By adopting a state-of-the-art second-order optimization algorithm, we train a Bayesian bilinear neural network with temporal attention.
By addressing the use of predictive distributions to analyze errors and uncertainties associated with the estimated parameters and model forecasts, we thoroughly compare our Bayesian model with traditional ML alternatives.
arXiv Detail & Related papers (2022-03-07T18:59:54Z) - Low-Rank Temporal Attention-Augmented Bilinear Network for financial
time-series forecasting [93.73198973454944]
Deep learning models have led to significant performance improvements in many problems coming from different domains, including prediction problems of financial time-series data.
The Temporal Attention-Augmented Bilinear network was recently proposed as an efficient and high-performing model for Limit Order Book time-series forecasting.
In this paper, we propose a low-rank tensor approximation of the model to further reduce the number of trainable parameters and increase its speed.
arXiv Detail & Related papers (2021-07-05T10:15:23Z) - Stock2Vec: A Hybrid Deep Learning Framework for Stock Market Prediction
with Representation Learning and Temporal Convolutional Network [71.25144476293507]
We have proposed to develop a global hybrid deep learning framework to predict the daily prices in the stock market.
With representation learning, we derived an embedding called Stock2Vec, which gives us insight for the relationship among different stocks.
Our hybrid framework integrates both advantages and achieves better performance on the stock price prediction task than several popular benchmarked models.
arXiv Detail & Related papers (2020-09-29T22:54:30Z) - Deep Stock Predictions [58.720142291102135]
We consider the design of a trading strategy that performs portfolio optimization using Long Short Term Memory (LSTM) neural networks.
We then customize the loss function used to train the LSTM to increase the profit earned.
We find the LSTM model with the customized loss function to have an improved performance in the training bot over a regressive baseline such as ARIMA.
arXiv Detail & Related papers (2020-06-08T23:37:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.