Towards a Complete Analysis of Langevin Monte Carlo: Beyond Poincar\'e
Inequality
- URL: http://arxiv.org/abs/2303.03589v2
- Date: Thu, 20 Jul 2023 17:15:01 GMT
- Title: Towards a Complete Analysis of Langevin Monte Carlo: Beyond Poincar\'e
Inequality
- Authors: Alireza Mousavi-Hosseini and Tyler Farghly and Ye He and Krishnakumar
Balasubramanian and Murat A. Erdogdu
- Abstract summary: This research program was initiated by Vempala and Wibisono, who established results under log-Sobolev inequalities.
Our results explicitly quantify the effect of the initializer on the performance of the Langevin Monte Carlo (LMC) algorithm.
- Score: 15.941909642134085
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Langevin diffusions are rapidly convergent under appropriate functional
inequality assumptions. Hence, it is natural to expect that with additional
smoothness conditions to handle the discretization errors, their
discretizations like the Langevin Monte Carlo (LMC) converge in a similar
fashion. This research program was initiated by Vempala and Wibisono (2019),
who established results under log-Sobolev inequalities. Chewi et al. (2022)
extended the results to handle the case of Poincar\'e inequalities. In this
paper, we go beyond Poincar\'e inequalities, and push this research program to
its limit. We do so by establishing upper and lower bounds for Langevin
diffusions and LMC under weak Poincar\'e inequalities that are satisfied by a
large class of densities including polynomially-decaying heavy-tailed densities
(i.e., Cauchy-type). Our results explicitly quantify the effect of the
initializer on the performance of the LMC algorithm. In particular, we show
that as the tail goes from sub-Gaussian, to sub-exponential, and finally to
Cauchy-like, the dependency on the initial error goes from being logarithmic,
to polynomial, and then finally to being exponential. This three-step phase
transition is in particular unavoidable as demonstrated by our lower bounds,
clearly defining the boundaries of LMC.
Related papers
- Error bounds for particle gradient descent, and extensions of the log-Sobolev and Talagrand inequalities [1.3124513975412255]
We show that for models satisfying a condition generalizing both the log-Sobolev and the Polyak--Lojasiewicz inequalities, the flow converges exponentially fast to the set of minimizers of the free energy.
We also generalize the Bakry--'Emery Theorem and show that the LSI/PLI generalization holds for models with strongly concave log-likelihoods.
arXiv Detail & Related papers (2024-03-04T12:57:26Z) - Taming under isoperimetry [0.0]
In this article we propose a Langevin-based scheme calledmathbfsTULA$ to sample from distributions with growing log.
We derive non-asymientKL and consequently consequently satisfy a Log-Sobolev inequality.
arXiv Detail & Related papers (2023-11-15T14:44:16Z) - When does Metropolized Hamiltonian Monte Carlo provably outperform
Metropolis-adjusted Langevin algorithm? [4.657614491309671]
We analyze the mixing time of Metropolized Hamiltonian Monte Carlo (HMC) with the leapfrog integrator.
We show that the joint distribution of the location and velocity variables of the discretization of the continuous HMC dynamics stays approximately invariant.
arXiv Detail & Related papers (2023-04-10T17:35:57Z) - Non-asymptotic analysis of Langevin-type Monte Carlo algorithms [0.0]
We study Langevin-type algorithms for sampling from Gibbs distributions with finite moduli of continuity not necessarily convergent to zero.
We show that the Langevin Monte Carlo algorithm can approximate Gibbs distributions with arbitrary accuracy if the potentials are dissipative and their gradients are uniformly continuous.
arXiv Detail & Related papers (2023-03-22T09:16:17Z) - High-Probability Bounds for Stochastic Optimization and Variational
Inequalities: the Case of Unbounded Variance [59.211456992422136]
We propose algorithms with high-probability convergence results under less restrictive assumptions.
These results justify the usage of the considered methods for solving problems that do not fit standard functional classes in optimization.
arXiv Detail & Related papers (2023-02-02T10:37:23Z) - Lassoed Tree Boosting [53.56229983630983]
We prove that a gradient boosted tree algorithm with early stopping faster than $n-1/4$ L2 convergence in the large nonparametric space of cadlag functions of bounded sectional variation.
Our convergence proofs are based on a novel, general theorem on early stopping with empirical loss minimizers of nested Donsker classes.
arXiv Detail & Related papers (2022-05-22T00:34:41Z) - Analysis of Langevin Monte Carlo from Poincaré to Log-Sobolev [25.18241929887685]
We provide the first convergence guarantees for the discrete-time Langevin Monte Carlo algorithm.
Unlike prior works, our results allow for weak smoothness and do not require convexity or dissipativity conditions.
arXiv Detail & Related papers (2021-12-23T15:56:33Z) - Mean-Square Analysis with An Application to Optimal Dimension Dependence
of Langevin Monte Carlo [60.785586069299356]
This work provides a general framework for the non-asymotic analysis of sampling error in 2-Wasserstein distance.
Our theoretical analysis is further validated by numerical experiments.
arXiv Detail & Related papers (2021-09-08T18:00:05Z) - A Fully Problem-Dependent Regret Lower Bound for Finite-Horizon MDPs [117.82903457289584]
We derive a novel problem-dependent lower-bound for regret in finite-horizon Markov Decision Processes (MDPs)
We show that our lower-bound is considerably smaller than in the general case and it does not scale with the minimum action gap at all.
We show that this last result is attainable (up to $poly(H)$ terms, where $H$ is the horizon) by providing a regret upper-bound based on policy gaps for an optimistic algorithm.
arXiv Detail & Related papers (2021-06-24T13:46:09Z) - Linear Last-iterate Convergence in Constrained Saddle-point Optimization [48.44657553192801]
We significantly expand the understanding of last-rate uniqueness for Optimistic Gradient Descent Ascent (OGDA) and Optimistic Multiplicative Weights Update (OMWU)
We show that when the equilibrium is unique, linear lastiterate convergence is achieved with a learning rate whose value is set to a universal constant.
We show that bilinear games over any polytope satisfy this condition and OGDA converges exponentially fast even without the unique equilibrium assumption.
arXiv Detail & Related papers (2020-06-16T20:53:04Z) - Approximation Schemes for ReLU Regression [80.33702497406632]
We consider the fundamental problem of ReLU regression.
The goal is to output the best fitting ReLU with respect to square loss given to draws from some unknown distribution.
arXiv Detail & Related papers (2020-05-26T16:26:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.