A Note on the Convergence of Mirrored Stein Variational Gradient Descent
under $(L_0,L_1)-$Smoothness Condition
- URL: http://arxiv.org/abs/2206.09709v1
- Date: Mon, 20 Jun 2022 11:04:18 GMT
- Title: A Note on the Convergence of Mirrored Stein Variational Gradient Descent
under $(L_0,L_1)-$Smoothness Condition
- Authors: Lukang Sun, Peter Richt\'arik
- Abstract summary: We establish a descent lemma for the population limit Mirrored Stein Variational Gradient Method(MSVGD)
This descent lemma does not rely on the path information of MSVGD but rather on a simple assumption for the mirrored distribution $nablaPsi_#piproptoexp(-V)$.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this note, we establish a descent lemma for the population limit Mirrored
Stein Variational Gradient Method~(MSVGD). This descent lemma does not rely on
the path information of MSVGD but rather on a simple assumption for the
mirrored distribution $\nabla\Psi_{\#}\pi\propto\exp(-V)$. Our analysis
demonstrates that MSVGD can be applied to a broader class of constrained
sampling problems with non-smooth $V$. We also investigate the complexity of
the population limit MSVGD in terms of dimension $d$.
Related papers
- Projection by Convolution: Optimal Sample Complexity for Reinforcement Learning in Continuous-Space MDPs [56.237917407785545]
We consider the problem of learning an $varepsilon$-optimal policy in a general class of continuous-space Markov decision processes (MDPs) having smooth Bellman operators.
Key to our solution is a novel projection technique based on ideas from harmonic analysis.
Our result bridges the gap between two popular but conflicting perspectives on continuous-space MDPs.
arXiv Detail & Related papers (2024-05-10T09:58:47Z) - Horizon-Free and Variance-Dependent Reinforcement Learning for Latent
Markov Decision Processes [62.90204655228324]
We study regret minimization for reinforcement learning (RL) in Latent Markov Decision Processes (LMDPs) with context in hindsight.
We design a novel model-based algorithmic framework which can be instantiated with both a model-optimistic and a value-optimistic solver.
arXiv Detail & Related papers (2022-10-20T21:32:01Z) - Improved Stein Variational Gradient Descent with Importance Weights [0.0]
Stein Variational Gradient Descent (SVGD) is a popular sampling algorithm used in various machine learning tasks.
We propose to enhance SVGD via the introduction of importance weights, which leads to a new method for which we coin the name $beta$-SVGD.
arXiv Detail & Related papers (2022-10-02T08:42:44Z) - Convergence of Stein Variational Gradient Descent under a Weaker
Smoothness Condition [0.0]
Stein Variational Gradient Descent (SVGD) is an important alternative to the Langevin-type algorithms for sampling from probability distributions.
In the existing theory of Langevin-type algorithms and SVGD, the potential function $V$ is often assumed to be $L$-smooth.
arXiv Detail & Related papers (2022-06-01T14:08:35Z) - Grassmann Stein Variational Gradient Descent [3.644031721554146]
Stein variational gradient descent (SVGD) is a deterministic particle inference algorithm that provides an efficient alternative to Markov chain Monte Carlo.
Recent developments have advocated projecting both the score function and the data onto real lines to sidestep this issue.
We propose Grassmann Stein variational gradient descent (GSVGD) as an alternative approach, which permits projections onto arbitrary dimensional subspaces.
arXiv Detail & Related papers (2022-02-07T15:36:03Z) - Complexity Analysis of Stein Variational Gradient Descent Under
Talagrand's Inequality T1 [12.848239550098697]
We study the complexity of Stein Variational Gradient Descent (SVGD), which is an algorithm to sample from $pi(x) propto exp(-Fx))
Our key assumption is that the target distribution satisfies the inequality's inequality T1.
arXiv Detail & Related papers (2021-06-06T09:51:32Z) - A New Framework for Variance-Reduced Hamiltonian Monte Carlo [88.84622104944503]
We propose a new framework of variance-reduced Hamiltonian Monte Carlo (HMC) methods for sampling from an $L$-smooth and $m$-strongly log-concave distribution.
We show that the unbiased gradient estimators, including SAGA and SVRG, based HMC methods achieve highest gradient efficiency with small batch size.
Experimental results on both synthetic and real-world benchmark data show that our new framework significantly outperforms the full gradient and gradient HMC approaches.
arXiv Detail & Related papers (2021-02-09T02:44:24Z) - Variance-Aware Confidence Set: Variance-Dependent Bound for Linear
Bandits and Horizon-Free Bound for Linear Mixture MDP [76.94328400919836]
We show how to construct variance-aware confidence sets for linear bandits and linear mixture Decision Process (MDP)
For linear bandits, we obtain an $widetildeO(mathrmpoly(d)sqrt1 + sum_i=1Ksigma_i2) regret bound, where $d is the feature dimension.
For linear mixture MDP, we obtain an $widetildeO(mathrmpoly(d)sqrtK)$ regret bound, where
arXiv Detail & Related papers (2021-01-29T18:57:52Z) - A Non-Asymptotic Analysis for Stein Variational Gradient Descent [44.30569261307296]
We provide a novel finite time analysis for the Stein Variational Gradient Descent algorithm.
We provide a descent lemma establishing that the algorithm decreases the objective at each iteration.
We also provide a convergence result of the finite particle system corresponding to the practical implementation of SVGD to its population version.
arXiv Detail & Related papers (2020-06-17T12:01:33Z) - A diffusion approach to Stein's method on Riemannian manifolds [65.36007959755302]
We exploit the relationship between the generator of a diffusion on $mathbf M$ with target invariant measure and its characterising Stein operator.
We derive Stein factors, which bound the solution to the Stein equation and its derivatives.
We imply that the bounds for $mathbb Rm$ remain valid when $mathbf M$ is a flat manifold.
arXiv Detail & Related papers (2020-03-25T17:03:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.