Applying Deep Learning to Calibrate Stochastic Volatility Models
- URL: http://arxiv.org/abs/2309.07843v2
- Date: Mon, 25 Sep 2023 09:34:54 GMT
- Title: Applying Deep Learning to Calibrate Stochastic Volatility Models
- Authors: Abir Sridi and Paul Bilokon
- Abstract summary: We develop a Differential Machine Learning (DML) approach to price vanilla European options.
The trained neural network dramatically reduces Heston calibration's time.
We compare their performance in reducing overfitting and improving the generalisation error.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Stochastic volatility models, where the volatility is a stochastic process,
can capture most of the essential stylized facts of implied volatility surfaces
and give more realistic dynamics of the volatility smile/skew. However, they
come with the significant issue that they take too long to calibrate.
Alternative calibration methods based on Deep Learning (DL) techniques have
been recently used to build fast and accurate solutions to the calibration
problem. Huge and Savine developed a Differential Machine Learning (DML)
approach, where Machine Learning models are trained on samples of not only
features and labels but also differentials of labels to features. The present
work aims to apply the DML technique to price vanilla European options (i.e.
the calibration instruments), more specifically, puts when the underlying asset
follows a Heston model and then calibrate the model on the trained network. DML
allows for fast training and accurate pricing. The trained neural network
dramatically reduces Heston calibration's computation time.
In this work, we also introduce different regularisation techniques, and we
apply them notably in the case of the DML. We compare their performance in
reducing overfitting and improving the generalisation error. The DML
performance is also compared to the classical DL (without differentiation) one
in the case of Feed-Forward Neural Networks. We show that the DML outperforms
the DL.
The complete code for our experiments is provided in the GitHub repository:
https://github.com/asridi/DML-Calibration-Heston-Model
Related papers
- What Do Learning Dynamics Reveal About Generalization in LLM Reasoning? [83.83230167222852]
We find that a model's generalization behavior can be effectively characterized by a training metric we call pre-memorization train accuracy.
By connecting a model's learning behavior to its generalization, pre-memorization train accuracy can guide targeted improvements to training strategies.
arXiv Detail & Related papers (2024-11-12T09:52:40Z) - CALICO: Confident Active Learning with Integrated Calibration [11.978551396144532]
We propose an AL framework that self-calibrates the confidence used for sample selection during the training process.
We show improved classification performance compared to a softmax-based classifier with fewer labeled samples.
arXiv Detail & Related papers (2024-07-02T15:05:19Z) - Towards Foundation Models for Scientific Machine Learning:
Characterizing Scaling and Transfer Behavior [32.74388989649232]
We study how pre-training could be used for scientific machine learning (SciML) applications.
We find that fine-tuning these models yields more performance gains as model size increases.
arXiv Detail & Related papers (2023-06-01T00:32:59Z) - Winner-Take-All Column Row Sampling for Memory Efficient Adaptation of Language Model [89.8764435351222]
We propose a new family of unbiased estimators called WTA-CRS, for matrix production with reduced variance.
Our work provides both theoretical and experimental evidence that, in the context of tuning transformers, our proposed estimators exhibit lower variance compared to existing ones.
arXiv Detail & Related papers (2023-05-24T15:52:08Z) - Bridging Precision and Confidence: A Train-Time Loss for Calibrating
Object Detection [58.789823426981044]
We propose a novel auxiliary loss formulation that aims to align the class confidence of bounding boxes with the accurateness of predictions.
Our results reveal that our train-time loss surpasses strong calibration baselines in reducing calibration error for both in and out-domain scenarios.
arXiv Detail & Related papers (2023-03-25T08:56:21Z) - Rethinking Soft Label in Label Distribution Learning Perspective [0.27719338074999533]
The primary goal of training in early convolutional neural networks (CNN) is the higher generalization performance of the model.
We investigated that performing label distribution learning (LDL) would enhance the model calibration in CNN training.
We performed several visualizations and analyses and witnessed several interesting behaviors in CNN training with the LDL.
arXiv Detail & Related papers (2023-01-31T06:47:19Z) - A Close Look into the Calibration of Pre-trained Language Models [56.998539510508515]
Pre-trained language models (PLMs) may fail in giving reliable estimates of their predictive uncertainty.
We study the dynamic change in PLMs' calibration performance in training.
We extend two recently proposed learnable methods that directly collect data to train models to have reasonable confidence estimations.
arXiv Detail & Related papers (2022-10-31T21:31:07Z) - Incremental Online Learning Algorithms Comparison for Gesture and Visual
Smart Sensors [68.8204255655161]
This paper compares four state-of-the-art algorithms in two real applications: gesture recognition based on accelerometer data and image classification.
Our results confirm these systems' reliability and the feasibility of deploying them in tiny-memory MCUs.
arXiv Detail & Related papers (2022-09-01T17:05:20Z) - Transfer Learning without Knowing: Reprogramming Black-box Machine
Learning Models with Scarce Data and Limited Resources [78.72922528736011]
We propose a novel approach, black-box adversarial reprogramming (BAR), that repurposes a well-trained black-box machine learning model.
Using zeroth order optimization and multi-label mapping techniques, BAR can reprogram a black-box ML model solely based on its input-output responses.
BAR outperforms state-of-the-art methods and yields comparable performance to the vanilla adversarial reprogramming method.
arXiv Detail & Related papers (2020-07-17T01:52:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.