Machine learning for option pricing: an empirical investigation of
network architectures
- URL: http://arxiv.org/abs/2307.07657v1
- Date: Fri, 14 Jul 2023 23:27:43 GMT
- Title: Machine learning for option pricing: an empirical investigation of
network architectures
- Authors: Laurens Van Mieghem, Antonis Papapantoleon, Jonas Papazoglou-Hennig
- Abstract summary: We investigate whether and how the choice of network architecture affects the accuracy and training time of a machine learning algorithm.
For option pricing problems, where we focus on the Black--Scholes and the Heston model, the generalized highway network architecture outperforms all other variants.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We consider the supervised learning problem of learning the price of an
option or the implied volatility given appropriate input data (model
parameters) and corresponding output data (option prices or implied
volatilities). The majority of articles in this literature considers a (plain)
feed forward neural network architecture in order to connect the neurons used
for learning the function mapping inputs to outputs. In this article, motivated
by methods in image classification and recent advances in machine learning
methods for PDEs, we investigate empirically whether and how the choice of
network architecture affects the accuracy and training time of a machine
learning algorithm. We find that for option pricing problems, where we focus on
the Black--Scholes and the Heston model, the generalized highway network
architecture outperforms all other variants, when considering the mean squared
error and the training time as criteria. Moreover, for the computation of the
implied volatility, after a necessary transformation, a variant of the DGM
architecture outperforms all other variants, when considering again the mean
squared error and the training time as criteria.
Related papers
- AI-Aided Kalman Filters [65.35350122917914]
The Kalman filter (KF) and its variants are among the most celebrated algorithms in signal processing.
Recent developments illustrate the possibility of fusing deep neural networks (DNNs) with classic Kalman-type filtering.
This article provides a tutorial-style overview of design approaches for incorporating AI in aiding KF-type algorithms.
arXiv Detail & Related papers (2024-10-16T06:47:53Z) - Transfer Learning on Multi-Dimensional Data: A Novel Approach to Neural Network-Based Surrogate Modeling [0.0]
Convolutional neural networks (CNNs) have gained popularity as the basis for such surrogate models.
We propose training a CNN surrogate model on a mixture of numerical solutions to both the $d$-dimensional problem and its ($d-1$)-dimensional approximation.
We demonstrate our approach on a multiphase flow test problem, using transfer learning to train a dense fully-convolutional encoder-decoder CNN on the two classes of data.
arXiv Detail & Related papers (2024-10-16T05:07:48Z) - Lessons on Datasets and Paradigms in Machine Learning for Symbolic Computation: A Case Study on CAD [0.0]
This study reports lessons on the importance of analysing datasets prior to machine learning.
We present results for a particular case study, the selection of variable ordering for cylindrical algebraic decomposition.
We introduce an augmentation technique for systems that allows us to balance and further augment the dataset.
arXiv Detail & Related papers (2024-01-24T10:12:43Z) - Systematic Evaluation of Deep Learning Models for Log-based Failure Prediction [3.3810628880631226]
This paper systematically investigates the combination of log data embedding strategies and Deep Learning (DL) types for failure prediction.
To that end, we propose a modular architecture to accommodate various configurations of embedding strategies and DL-based encoders.
Using the F1 score metric, our results show that the best overall performing configuration is a CNN-based encoder with Logkey2vec.
arXiv Detail & Related papers (2023-03-13T16:04:14Z) - VCNet: A self-explaining model for realistic counterfactual generation [52.77024349608834]
Counterfactual explanation is a class of methods to make local explanations of machine learning decisions.
We present VCNet-Variational Counter Net, a model architecture that combines a predictor and a counterfactual generator.
We show that VCNet is able to both generate predictions, and to generate counterfactual explanations without having to solve another minimisation problem.
arXiv Detail & Related papers (2022-12-21T08:45:32Z) - Deep invariant networks with differentiable augmentation layers [87.22033101185201]
Methods for learning data augmentation policies require held-out data and are based on bilevel optimization problems.
We show that our approach is easier and faster to train than modern automatic data augmentation techniques.
arXiv Detail & Related papers (2022-02-04T14:12:31Z) - Adaptive Anomaly Detection for Internet of Things in Hierarchical Edge
Computing: A Contextual-Bandit Approach [81.5261621619557]
We propose an adaptive anomaly detection scheme with hierarchical edge computing (HEC)
We first construct multiple anomaly detection DNN models with increasing complexity, and associate each of them to a corresponding HEC layer.
Then, we design an adaptive model selection scheme that is formulated as a contextual-bandit problem and solved by using a reinforcement learning policy network.
arXiv Detail & Related papers (2021-08-09T08:45:47Z) - TELESTO: A Graph Neural Network Model for Anomaly Classification in
Cloud Services [77.454688257702]
Machine learning (ML) and artificial intelligence (AI) are applied on IT system operation and maintenance.
One direction aims at the recognition of re-occurring anomaly types to enable remediation automation.
We propose a method that is invariant to dimensionality changes of given data.
arXiv Detail & Related papers (2021-02-25T14:24:49Z) - Using Data Assimilation to Train a Hybrid Forecast System that Combines
Machine-Learning and Knowledge-Based Components [52.77024349608834]
We consider the problem of data-assisted forecasting of chaotic dynamical systems when the available data is noisy partial measurements.
We show that by using partial measurements of the state of the dynamical system, we can train a machine learning model to improve predictions made by an imperfect knowledge-based model.
arXiv Detail & Related papers (2021-02-15T19:56:48Z) - More Is More -- Narrowing the Generalization Gap by Adding
Classification Heads [8.883733362171032]
We introduce an architecture enhancement for existing neural network models based on input transformations, termed 'TransNet'
Our model can be employed during training time only and then pruned for prediction, resulting in an equivalent architecture to the base model.
arXiv Detail & Related papers (2021-02-09T16:30:33Z) - Learning to Rank Learning Curves [15.976034696758148]
We present a new method that saves computational budget by terminating poor configurations early on in the training.
We show that our model is able to effectively rank learning curves without having to observe many or very long learning curves.
arXiv Detail & Related papers (2020-06-05T10:49:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.