Deep combinatorial optimisation for optimal stopping time problems :
application to swing options pricing
- URL: http://arxiv.org/abs/2001.11247v2
- Date: Fri, 8 Jan 2021 14:55:16 GMT
- Title: Deep combinatorial optimisation for optimal stopping time problems :
application to swing options pricing
- Authors: Thomas Deschatre and Joseph Mikael
- Abstract summary: A new method for computation control based on neural networks and using randomisation of discrete random variables is proposed and applied to optimal stopping time problems.
The proposed algorithm succeeds in pricing high dimensional American and swing options in a reasonable time, which is not possible with classical algorithms.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: A new method for stochastic control based on neural networks and using
randomisation of discrete random variables is proposed and applied to optimal
stopping time problems. The method models directly the policy and does not need
the derivation of a dynamic programming principle nor a backward stochastic
differential equation. Unlike continuous optimization where automatic
differentiation is used directly, we propose a likelihood ratio method for
gradient computation. Numerical tests are done on the pricing of American and
swing options. The proposed algorithm succeeds in pricing high dimensional
American and swing options in a reasonable computation time, which is not
possible with classical algorithms.
Related papers
- Regret-Optimal Model-Free Reinforcement Learning for Discounted MDPs
with Short Burn-In Time [13.545356254920584]
We introduce a model-free algorithm that employs variance reduction and a novel technique that switches the execution policy in a slow-yet-adaptive manner.
This is the first regret-optimal model-free algorithm in the discounted setting, with the additional benefit of a low burn-in time.
arXiv Detail & Related papers (2023-05-24T20:22:43Z) - Accelerated First-Order Optimization under Nonlinear Constraints [73.2273449996098]
We exploit between first-order algorithms for constrained optimization and non-smooth systems to design a new class of accelerated first-order algorithms.
An important property of these algorithms is that constraints are expressed in terms of velocities instead of sparse variables.
arXiv Detail & Related papers (2023-02-01T08:50:48Z) - Maximum-Likelihood Inverse Reinforcement Learning with Finite-Time
Guarantees [56.848265937921354]
Inverse reinforcement learning (IRL) aims to recover the reward function and the associated optimal policy.
Many algorithms for IRL have an inherently nested structure.
We develop a novel single-loop algorithm for IRL that does not compromise reward estimation accuracy.
arXiv Detail & Related papers (2022-10-04T17:13:45Z) - Latency considerations for stochastic optimizers in variational quantum
algorithms [0.02294014185517203]
Variational quantum algorithms, which have risen to prominence in the noisy intermediate noisy quantum-scale setting, require the implementation of hardware.
To date, most research has employed algorithms based on the gradient iteration as the classical iteration.
In this work we propose instead using optimization algorithms that yield processes emulating the dynamics of classical deterministic algorithms.
arXiv Detail & Related papers (2022-01-31T18:51:24Z) - Momentum Accelerates the Convergence of Stochastic AUPRC Maximization [80.8226518642952]
We study optimization of areas under precision-recall curves (AUPRC), which is widely used for imbalanced tasks.
We develop novel momentum methods with a better iteration of $O (1/epsilon4)$ for finding an $epsilon$stationary solution.
We also design a novel family of adaptive methods with the same complexity of $O (1/epsilon4)$, which enjoy faster convergence in practice.
arXiv Detail & Related papers (2021-07-02T16:21:52Z) - An Adaptive Stochastic Sequential Quadratic Programming with
Differentiable Exact Augmented Lagrangians [17.9230793188835]
We consider the problem of solving nonlinear optimization programs with objective and deterministic equality.
We propose a sequential quadratic programming (SQP) that uses a differentiable exact augmented Lagrangian as the merit function.
The proposed algorithm is the first SQP that allows a line search procedure and the first line search procedure.
arXiv Detail & Related papers (2021-02-10T08:40:55Z) - Divide and Learn: A Divide and Conquer Approach for Predict+Optimize [50.03608569227359]
The predict+optimize problem combines machine learning ofproblem coefficients with a optimization prob-lem that uses the predicted coefficients.
We show how to directlyexpress the loss of the optimization problem in terms of thepredicted coefficients as a piece-wise linear function.
We propose a novel divide and algorithm to tackle optimization problems without this restriction and predict itscoefficients using the optimization loss.
arXiv Detail & Related papers (2020-12-04T00:26:56Z) - Variance-Reduced Off-Policy Memory-Efficient Policy Search [61.23789485979057]
Off-policy policy optimization is a challenging problem in reinforcement learning.
Off-policy algorithms are memory-efficient and capable of learning from off-policy samples.
arXiv Detail & Related papers (2020-09-14T16:22:46Z) - Stochastic Adaptive Line Search for Differentially Private Optimization [6.281099620056346]
The performance of private gradient-based optimization algorithms is highly dependent on the choice step size (or learning rate)
We introduce a variant of classic non-trivial line search algorithm that adjusts the privacy gradient according to the reliability of noisy gradient.
We show that the adaptively chosen step sizes allow the proposed algorithm to efficiently use the privacy budget and show competitive performance against existing private gradients.
arXiv Detail & Related papers (2020-08-18T15:18:47Z) - Sequential Quadratic Optimization for Nonlinear Equality Constrained
Stochastic Optimization [10.017195276758454]
It is assumed in this setting that it is intractable to compute objective function and derivative values explicitly.
An algorithm is proposed for the deterministic setting that is modeled after a state-of-the-art line-search SQP algorithm.
The results of numerical experiments demonstrate the practical performance of our proposed techniques.
arXiv Detail & Related papers (2020-07-20T23:04:26Z) - Private Stochastic Convex Optimization: Optimal Rates in Linear Time [74.47681868973598]
We study the problem of minimizing the population loss given i.i.d. samples from a distribution over convex loss functions.
A recent work of Bassily et al. has established the optimal bound on the excess population loss achievable given $n$ samples.
We describe two new techniques for deriving convex optimization algorithms both achieving the optimal bound on excess loss and using $O(minn, n2/d)$ gradient computations.
arXiv Detail & Related papers (2020-05-10T19:52:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.