Generative Modelling of L\'{e}vy Area for High Order SDE Simulation
- URL: http://arxiv.org/abs/2308.02452v1
- Date: Fri, 4 Aug 2023 16:38:37 GMT
- Title: Generative Modelling of L\'{e}vy Area for High Order SDE Simulation
- Authors: Andra\v{z} Jelin\v{c}i\v{c}, Jiajie Tao, William F. Turner, Thomas
Cass, James Foster, Hao Ni
- Abstract summary: L'evyGAN is a deep-learning model for generating approximate samples of L'evy area conditional on a Brownian increment.
We show that L'evyGAN exhibits state-of-the-art performance across several metrics which measure both the joint and marginal distributions.
- Score: 5.9535699822923
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: It is well known that, when numerically simulating solutions to SDEs,
achieving a strong convergence rate better than O(\sqrt{h}) (where h is the
step size) requires the use of certain iterated integrals of Brownian motion,
commonly referred to as its "L\'{e}vy areas". However, these stochastic
integrals are difficult to simulate due to their non-Gaussian nature and for a
d-dimensional Brownian motion with d > 2, no fast almost-exact sampling
algorithm is known.
In this paper, we propose L\'{e}vyGAN, a deep-learning-based model for
generating approximate samples of L\'{e}vy area conditional on a Brownian
increment. Due to our "Bridge-flipping" operation, the output samples match all
joint and conditional odd moments exactly. Our generator employs a tailored
GNN-inspired architecture, which enforces the correct dependency structure
between the output distribution and the conditioning variable. Furthermore, we
incorporate a mathematically principled characteristic-function based
discriminator. Lastly, we introduce a novel training mechanism termed
"Chen-training", which circumvents the need for expensive-to-generate training
data-sets. This new training procedure is underpinned by our two main
theoretical results.
For 4-dimensional Brownian motion, we show that L\'{e}vyGAN exhibits
state-of-the-art performance across several metrics which measure both the
joint and marginal distributions. We conclude with a numerical experiment on
the log-Heston model, a popular SDE in mathematical finance, demonstrating that
high-quality synthetic L\'{e}vy area can lead to high order weak convergence
and variance reduction when using multilevel Monte Carlo (MLMC).
Related papers
- Bellman Diffusion: Generative Modeling as Learning a Linear Operator in the Distribution Space [72.52365911990935]
We introduce Bellman Diffusion, a novel DGM framework that maintains linearity in MDPs through gradient and scalar field modeling.
Our results show that Bellman Diffusion achieves accurate field estimations and is a capable image generator, converging 1.5x faster than the traditional histogram-based baseline in distributional RL tasks.
arXiv Detail & Related papers (2024-10-02T17:53:23Z) - HJ-sampler: A Bayesian sampler for inverse problems of a stochastic process by leveraging Hamilton-Jacobi PDEs and score-based generative models [1.949927790632678]
This paper builds on the log transform known as the Cole-Hopf transform in Brownian motion contexts.
We develop a new algorithm, named the HJ-sampler, for inference for the inverse problem of a differential equation with given terminal observations.
arXiv Detail & Related papers (2024-09-15T05:30:54Z) - Faster Sampling via Stochastic Gradient Proximal Sampler [28.422547264326468]
Proximal samplers (SPS) for sampling from non-log-concave distributions are studied.
We show that the convergence to the target distribution can be guaranteed as long as the algorithm trajectory is bounded.
We provide two implementable variants based on Langevin dynamics (SGLD) and Langevin-MALA, giving rise to SPS-SGLD and SPS-MALA.
arXiv Detail & Related papers (2024-05-27T00:53:18Z) - On the Trajectory Regularity of ODE-based Diffusion Sampling [79.17334230868693]
Diffusion-based generative models use differential equations to establish a smooth connection between a complex data distribution and a tractable prior distribution.
In this paper, we identify several intriguing trajectory properties in the ODE-based sampling process of diffusion models.
arXiv Detail & Related papers (2024-05-18T15:59:41Z) - Single-seed generation of Brownian paths and integrals for adaptive and high order SDE solvers [9.879281328700577]
We present two applications of adaptive high order solvers enabled by our new VBT algorithm.
Using adaptive solvers to simulate a high-volatility CIR model, we achieve more than twice the convergence order of constant stepping.
We apply an adaptive third order underdamped or kinetic Langevin solver to an MCMC problem, while using only a tenth of its function evaluations.
arXiv Detail & Related papers (2024-05-10T13:16:23Z) - Gaussian Mixture Solvers for Diffusion Models [84.83349474361204]
We introduce a novel class of SDE-based solvers called GMS for diffusion models.
Our solver outperforms numerous SDE-based solvers in terms of sample quality in image generation and stroke-based synthesis.
arXiv Detail & Related papers (2023-11-02T02:05:38Z) - Instability and Local Minima in GAN Training with Kernel Discriminators [20.362912591032636]
Generative Adversarial Networks (GANs) are a widely-used tool for generative modeling of complex data.
Despite their empirical success, the training of GANs is not fully understood due to the min-max optimization of the generator and discriminator.
This paper analyzes these joint dynamics when the true samples, as well as the generated samples, are discrete, finite sets, and the discriminator is kernel-based.
arXiv Detail & Related papers (2022-08-21T18:03:06Z) - Improved Convergence Rate of Stochastic Gradient Langevin Dynamics with
Variance Reduction and its Application to Optimization [50.83356836818667]
gradient Langevin Dynamics is one of the most fundamental algorithms to solve non-eps optimization problems.
In this paper, we show two variants of this kind, namely the Variance Reduced Langevin Dynamics and the Recursive Gradient Langevin Dynamics.
arXiv Detail & Related papers (2022-03-30T11:39:00Z) - DASHA: Distributed Nonconvex Optimization with Communication
Compression, Optimal Oracle Complexity, and No Client Synchronization [77.34726150561087]
We develop and analyze DASHA: a new family of methods for noneps distributed optimization problems.
Unlike MARINA, the new methods DASHA, DASHA-MVR send compressed vectors only and never synchronize the nodes, which makes them more practical for learning.
arXiv Detail & Related papers (2022-02-02T20:10:40Z) - Mean-Square Analysis with An Application to Optimal Dimension Dependence
of Langevin Monte Carlo [60.785586069299356]
This work provides a general framework for the non-asymotic analysis of sampling error in 2-Wasserstein distance.
Our theoretical analysis is further validated by numerical experiments.
arXiv Detail & Related papers (2021-09-08T18:00:05Z) - Dual Stochastic Natural Gradient Descent and convergence of interior
half-space gradient approximations [0.0]
Multinomial logistic regression (MLR) is widely used in statistics and machine learning.
gradient descent (SGD) is the most common approach for determining the parameters of a MLR model in big data scenarios.
arXiv Detail & Related papers (2020-01-19T00:53:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.