Riemannian Langevin Algorithm for Solving Semidefinite Programs
- URL: http://arxiv.org/abs/2010.11176v6
- Date: Mon, 19 Jun 2023 16:16:18 GMT
- Title: Riemannian Langevin Algorithm for Solving Semidefinite Programs
- Authors: Mufan Bill Li, Murat A. Erdogdu
- Abstract summary: We propose a Langevin-based algorithm for non- optimization and sampling on a product manifold of spheres.
We show that the proposed Langevin algorithm achieves $epsilon accuracy with high probability.
- Score: 9.340611077939828
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose a Langevin diffusion-based algorithm for non-convex optimization
and sampling on a product manifold of spheres. Under a logarithmic Sobolev
inequality, we establish a guarantee for finite iteration convergence to the
Gibbs distribution in terms of Kullback--Leibler divergence. We show that with
an appropriate temperature choice, the suboptimality gap to the global minimum
is guaranteed to be arbitrarily small with high probability.
As an application, we consider the Burer--Monteiro approach for solving a
semidefinite program (SDP) with diagonal constraints, and analyze the proposed
Langevin algorithm for optimizing the non-convex objective. In particular, we
establish a logarithmic Sobolev inequality for the Burer--Monteiro problem when
there are no spurious local minima, but under the presence saddle points.
Combining the results, we then provide a global optimality guarantee for the
SDP and the Max-Cut problem. More precisely, we show that the Langevin
algorithm achieves $\epsilon$ accuracy with high probability in
$\widetilde{\Omega}( \epsilon^{-5} )$ iterations.
Related papers
- Faster Sampling via Stochastic Gradient Proximal Sampler [28.422547264326468]
Proximal samplers (SPS) for sampling from non-log-concave distributions are studied.
We show that the convergence to the target distribution can be guaranteed as long as the algorithm trajectory is bounded.
We provide two implementable variants based on Langevin dynamics (SGLD) and Langevin-MALA, giving rise to SPS-SGLD and SPS-MALA.
arXiv Detail & Related papers (2024-05-27T00:53:18Z) - Symmetric Mean-field Langevin Dynamics for Distributional Minimax
Problems [78.96969465641024]
We extend mean-field Langevin dynamics to minimax optimization over probability distributions for the first time with symmetric and provably convergent updates.
We also study time and particle discretization regimes and prove a new uniform-in-time propagation of chaos result.
arXiv Detail & Related papers (2023-12-02T13:01:29Z) - Decentralized Riemannian Algorithm for Nonconvex Minimax Problems [82.50374560598493]
The minimax algorithms for neural networks have been developed to solve many problems.
In this paper, we propose two types of minimax algorithms.
For the setting, we propose DRSGDA and prove that our method achieves a gradient.
arXiv Detail & Related papers (2023-02-08T01:42:45Z) - Kinetic Langevin MCMC Sampling Without Gradient Lipschitz Continuity --
the Strongly Convex Case [0.0]
We consider sampling from log concave distributions in Hamiltonian setting, without assuming that the objective is globally Lipschitz.
We propose two algorithms based on polygonal gradient (tamed) Euler schemes, to sample from a target measure, and provide non-asymptotic 2-Wasserstein distance bounds between the law of the process of each algorithm and the target measure.
arXiv Detail & Related papers (2023-01-19T12:32:41Z) - Optimal policy evaluation using kernel-based temporal difference methods [78.83926562536791]
We use kernel Hilbert spaces for estimating the value function of an infinite-horizon discounted Markov reward process.
We derive a non-asymptotic upper bound on the error with explicit dependence on the eigenvalues of the associated kernel operator.
We prove minimax lower bounds over sub-classes of MRPs.
arXiv Detail & Related papers (2021-09-24T14:48:20Z) - Mean-Square Analysis with An Application to Optimal Dimension Dependence
of Langevin Monte Carlo [60.785586069299356]
This work provides a general framework for the non-asymotic analysis of sampling error in 2-Wasserstein distance.
Our theoretical analysis is further validated by numerical experiments.
arXiv Detail & Related papers (2021-09-08T18:00:05Z) - Projected Stochastic Gradient Langevin Algorithms for Constrained
Sampling and Non-Convex Learning [0.0]
Langevin algorithms are methods with additive noise.
Langevin algorithms have been used for decades in chain Carlo (Milon)
For learning, we show that it is that it is that it is that it is that it is that it is that it is that it is that it is that it is that it is that it is that it is that it is that it is that it is that it is that it is that it is that it is that it is that it is that it is that it is that it is that it is that it is that it is that it is that it is that it is that
arXiv Detail & Related papers (2020-12-22T16:19:20Z) - Faster Convergence of Stochastic Gradient Langevin Dynamics for
Non-Log-Concave Sampling [110.88857917726276]
We provide a new convergence analysis of gradient Langevin dynamics (SGLD) for sampling from a class of distributions that can be non-log-concave.
At the core of our approach is a novel conductance analysis of SGLD using an auxiliary time-reversible Markov Chain.
arXiv Detail & Related papers (2020-10-19T15:23:18Z) - Primal Dual Interpretation of the Proximal Stochastic Gradient Langevin
Algorithm [11.80267432402723]
We consider the task of sampling with respect to a log concave probability distribution.
The target distribution can be seen as a minimizer of the Kullback-Leibler divergence defined on the Wasserstein space.
We show that if the potential is strongly convex, the complexity of PSGLA is $O (1/varepsilon2)$ in terms of the 2-Wasserstein distance.
arXiv Detail & Related papers (2020-06-16T15:57:09Z) - Stochastic Saddle-Point Optimization for Wasserstein Barycenters [69.68068088508505]
We consider the populationimation barycenter problem for random probability measures supported on a finite set of points and generated by an online stream of data.
We employ the structure of the problem and obtain a convex-concave saddle-point reformulation of this problem.
In the setting when the distribution of random probability measures is discrete, we propose an optimization algorithm and estimate its complexity.
arXiv Detail & Related papers (2020-06-11T19:40:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.