Taming under isoperimetry
- URL: http://arxiv.org/abs/2311.09003v1
- Date: Wed, 15 Nov 2023 14:44:16 GMT
- Title: Taming under isoperimetry
- Authors: Iosif Lytras and Sotirios Sabanis
- Abstract summary: In this article we propose a Langevin-based scheme calledmathbfsTULA$ to sample from distributions with growing log.
We derive non-asymientKL and consequently consequently satisfy a Log-Sobolev inequality.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this article we propose a novel taming Langevin-based scheme called
$\mathbf{sTULA}$ to sample from distributions with superlinearly growing
log-gradient which also satisfy a Log-Sobolev inequality. We derive
non-asymptotic convergence bounds in $KL$ and consequently total variation and
Wasserstein-$2$ distance from the target measure. Non-asymptotic convergence
guarantees are provided for the performance of the new algorithm as an
optimizer. Finally, some theoretical results on isoperimertic inequalities for
distributions with superlinearly growing gradients are provided. Key findings
are a Log-Sobolev inequality with constant independent of the dimension, in the
presence of a higher order regularization and a Poincare inequality with
constant independent of temperature and dimension under a novel non-convex
theoretical framework.
Related papers
- Fast Convergence of $Φ$-Divergence Along the Unadjusted Langevin Algorithm and Proximal Sampler [14.34147140416535]
We study the mixing time of two popular discrete time Markov chains in continuous space.
We show that any $Phi$-divergence arising from a twice-differentiable strictly convex function $Phi$ converges to $0$ exponentially fast along these Markov chains.
arXiv Detail & Related papers (2024-10-14T16:41:45Z) - Provable Convergence and Limitations of Geometric Tempering for Langevin Dynamics [8.683011785637824]
Geometric tempering is a popular approach to sampling from challenging multi-modal probability distributions.
In this paper, we theoretically investigate the soundness of this approach when the sampling algorithm is Langevin dynamics.
Our results indicate that geometric tempering may not help, and can even be harmful for convergence.
arXiv Detail & Related papers (2024-10-13T02:24:31Z) - Non-asymptotic estimates for accelerated high order Langevin Monte Carlo algorithms [8.058385158111207]
We propose two new algorithms, namely aHOLA and aHOLLA, to sample from high-dimensional target distributions.
We establish non-asymptotic convergence rates for aHOLA in Wasserstein-1 and Wasserstein-2 distributions.
arXiv Detail & Related papers (2024-05-09T11:12:03Z) - Symmetric Mean-field Langevin Dynamics for Distributional Minimax
Problems [78.96969465641024]
We extend mean-field Langevin dynamics to minimax optimization over probability distributions for the first time with symmetric and provably convergent updates.
We also study time and particle discretization regimes and prove a new uniform-in-time propagation of chaos result.
arXiv Detail & Related papers (2023-12-02T13:01:29Z) - Convergence and concentration properties of constant step-size SGD
through Markov chains [0.0]
We consider the optimization of a smooth and strongly convex objective using constant step-size gradient descent (SGD)
We show that, for unbiased gradient estimates with mildly controlled variance, the iteration converges to an invariant distribution in total variation distance.
All our results are non-asymptotic and their consequences are discussed through a few applications.
arXiv Detail & Related papers (2023-06-20T12:36:28Z) - High-Probability Bounds for Stochastic Optimization and Variational
Inequalities: the Case of Unbounded Variance [59.211456992422136]
We propose algorithms with high-probability convergence results under less restrictive assumptions.
These results justify the usage of the considered methods for solving problems that do not fit standard functional classes in optimization.
arXiv Detail & Related papers (2023-02-02T10:37:23Z) - Mean-Square Analysis with An Application to Optimal Dimension Dependence
of Langevin Monte Carlo [60.785586069299356]
This work provides a general framework for the non-asymotic analysis of sampling error in 2-Wasserstein distance.
Our theoretical analysis is further validated by numerical experiments.
arXiv Detail & Related papers (2021-09-08T18:00:05Z) - Heavy-tailed Streaming Statistical Estimation [58.70341336199497]
We consider the task of heavy-tailed statistical estimation given streaming $p$ samples.
We design a clipped gradient descent and provide an improved analysis under a more nuanced condition on the noise of gradients.
arXiv Detail & Related papers (2021-08-25T21:30:27Z) - Spectral clustering under degree heterogeneity: a case for the random
walk Laplacian [83.79286663107845]
This paper shows that graph spectral embedding using the random walk Laplacian produces vector representations which are completely corrected for node degree.
In the special case of a degree-corrected block model, the embedding concentrates about K distinct points, representing communities.
arXiv Detail & Related papers (2021-05-03T16:36:27Z) - Faster Convergence of Stochastic Gradient Langevin Dynamics for
Non-Log-Concave Sampling [110.88857917726276]
We provide a new convergence analysis of gradient Langevin dynamics (SGLD) for sampling from a class of distributions that can be non-log-concave.
At the core of our approach is a novel conductance analysis of SGLD using an auxiliary time-reversible Markov Chain.
arXiv Detail & Related papers (2020-10-19T15:23:18Z) - Generalized Bayesian Cram\'{e}r-Rao Inequality via Information Geometry
of Relative $\alpha$-Entropy [17.746238062801293]
relative $alpha$-entropy is the R'enyi analog of relative entropy.
Recent information geometric investigations on this quantity have enabled the generalization of the Cram'er-Rao inequality.
We show that in the limiting case when the entropy order approaches unity, this framework reduces to the conventional Bayesian Cram'er-Rao inequality.
arXiv Detail & Related papers (2020-02-11T23:38:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.