Asymptotic bias of inexact Markov Chain Monte Carlo methods in high
dimension
- URL: http://arxiv.org/abs/2108.00682v2
- Date: Wed, 12 Apr 2023 09:25:54 GMT
- Title: Asymptotic bias of inexact Markov Chain Monte Carlo methods in high
dimension
- Authors: Alain Oliviero Durmus and Andreas Eberle
- Abstract summary: Examples include the unadjusted Langevin (ULA) and unadjusted Hamiltonian Monte Carlo (uHMC)
We show that for both ULA and uHMC, the bias depends on key quantities related to the target distribution or the stationary probability measure of the scheme.
- Score: 0.7614628596146599
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Inexact Markov Chain Monte Carlo methods rely on Markov chains that do not
exactly preserve the target distribution. Examples include the unadjusted
Langevin algorithm (ULA) and unadjusted Hamiltonian Monte Carlo (uHMC). This
paper establishes bounds on Wasserstein distances between the invariant
probability measures of inexact MCMC methods and their target distributions
with a focus on understanding the precise dependence of this asymptotic bias on
both dimension and discretization step size. Assuming Wasserstein bounds on the
convergence to equilibrium of either the exact or the approximate dynamics, we
show that for both ULA and uHMC, the asymptotic bias depends on key quantities
related to the target distribution or the stationary probability measure of the
scheme. As a corollary, we conclude that for models with a limited amount of
interactions such as mean-field models, finite range graphical models, and
perturbations thereof, the asymptotic bias has a similar dependence on the step
size and the dimension as for product measures.
Related papers
- Randomized Physics-Informed Machine Learning for Uncertainty
Quantification in High-Dimensional Inverse Problems [49.1574468325115]
We propose a physics-informed machine learning method for uncertainty quantification in high-dimensional inverse problems.
We show analytically and through comparison with Hamiltonian Monte Carlo that the rPICKLE posterior converges to the true posterior given by the Bayes rule.
arXiv Detail & Related papers (2023-12-11T07:33:16Z) - Unbiased Kinetic Langevin Monte Carlo with Inexact Gradients [0.8749675983608172]
We present an unbiased method for posterior means based on kinetic Langevin dynamics.
Our proposed estimator is unbiased, attains finite variance, and satisfies a central limit theorem.
Our results demonstrate that in large-scale applications, the unbiased algorithm we present can be 2-3 orders of magnitude more efficient than the gold-standard" randomized Hamiltonian Monte Carlo.
arXiv Detail & Related papers (2023-11-08T21:19:52Z) - Adaptive Annealed Importance Sampling with Constant Rate Progress [68.8204255655161]
Annealed Importance Sampling (AIS) synthesizes weighted samples from an intractable distribution.
We propose the Constant Rate AIS algorithm and its efficient implementation for $alpha$-divergences.
arXiv Detail & Related papers (2023-06-27T08:15:28Z) - Efficient CDF Approximations for Normalizing Flows [64.60846767084877]
We build upon the diffeomorphic properties of normalizing flows to estimate the cumulative distribution function (CDF) over a closed region.
Our experiments on popular flow architectures and UCI datasets show a marked improvement in sample efficiency as compared to traditional estimators.
arXiv Detail & Related papers (2022-02-23T06:11:49Z) - Bounding Wasserstein distance with couplings [26.17941985324059]
We propose estimators based on couplings of Markov chains to assess the quality of suchally biased sampling methods.
We establish theoretical guarantees for our upper bounds and show that our estimators can remain effective in high dimensions.
arXiv Detail & Related papers (2021-12-06T16:38:21Z) - Mean-Square Analysis with An Application to Optimal Dimension Dependence
of Langevin Monte Carlo [60.785586069299356]
This work provides a general framework for the non-asymotic analysis of sampling error in 2-Wasserstein distance.
Our theoretical analysis is further validated by numerical experiments.
arXiv Detail & Related papers (2021-09-08T18:00:05Z) - Machine Learning and Variational Algorithms for Lattice Field Theory [1.198562319289569]
In lattice quantum field theory studies, parameters defining the lattice theory must be tuned toward criticality to access continuum physics.
We introduce an approach to "deform" Monte Carlo estimators based on contour deformations applied to the domain of the path integral.
We demonstrate that flow-based MCMC can mitigate critical slowing down and observifolds can exponentially reduce variance in proof-of-principle applications.
arXiv Detail & Related papers (2021-06-03T16:37:05Z) - Spectral clustering under degree heterogeneity: a case for the random
walk Laplacian [83.79286663107845]
This paper shows that graph spectral embedding using the random walk Laplacian produces vector representations which are completely corrected for node degree.
In the special case of a degree-corrected block model, the embedding concentrates about K distinct points, representing communities.
arXiv Detail & Related papers (2021-05-03T16:36:27Z) - Signatures of Chaos in Non-integrable Models of Quantum Field Theory [0.0]
We study signatures of quantum chaos in (1+1)D Quantum Field Theory (QFT) models.
We focus on the double sine-Gordon, also studying the massive sine-Gordon and $phi4$ model.
arXiv Detail & Related papers (2020-12-15T18:56:20Z) - On Linear Stochastic Approximation: Fine-grained Polyak-Ruppert and
Non-Asymptotic Concentration [115.1954841020189]
We study the inequality and non-asymptotic properties of approximation procedures with Polyak-Ruppert averaging.
We prove a central limit theorem (CLT) for the averaged iterates with fixed step size and number of iterations going to infinity.
arXiv Detail & Related papers (2020-04-09T17:54:18Z) - Explicit Mean-Square Error Bounds for Monte-Carlo and Linear Stochastic
Approximation [4.817429789586127]
It is not possible to obtain a Hoeffding bound on the error sequence, even when the underlying Markov chain is reversible and geometrically ergodic.
It is shown that mean square error achieves the optimal rate of $O(1/n)$, subject to conditions on the step-size sequence.
arXiv Detail & Related papers (2020-02-07T01:52:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.