A stochastic Stein Variational Newton method
- URL: http://arxiv.org/abs/2204.09039v1
- Date: Tue, 19 Apr 2022 17:57:36 GMT
- Title: A stochastic Stein Variational Newton method
- Authors: Alex Leviyev, Joshua Chen, Yifei Wang, Omar Ghattas, Aaron Zimmerman
- Abstract summary: We show that Stein variational Newton (sSVN) is a promising approach to accelerating high-precision Bayesian inference tasks.
We demonstrate the effectiveness of our algorithm on a difficult class of test problems -- the Hybrid Rosenbrock density -- and show that sSVN converges using three orders of fewer magnitude evaluations of the log likelihood.
- Score: 7.272730677575111
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Stein variational gradient descent (SVGD) is a general-purpose
optimization-based sampling algorithm that has recently exploded in popularity,
but is limited by two issues: it is known to produce biased samples, and it can
be slow to converge on complicated distributions. A recently proposed
stochastic variant of SVGD (sSVGD) addresses the first issue, producing
unbiased samples by incorporating a special noise into the SVGD dynamics such
that asymptotic convergence is guaranteed. Meanwhile, Stein variational Newton
(SVN), a Newton-like extension of SVGD, dramatically accelerates the
convergence of SVGD by incorporating Hessian information into the dynamics, but
also produces biased samples. In this paper we derive, and provide a practical
implementation of, a stochastic variant of SVN (sSVN) which is both
asymptotically correct and converges rapidly. We demonstrate the effectiveness
of our algorithm on a difficult class of test problems -- the Hybrid Rosenbrock
density -- and show that sSVN converges using three orders of magnitude fewer
gradient evaluations of the log likelihood than its stochastic SVGD
counterpart. Our results show that sSVN is a promising approach to accelerating
high-precision Bayesian inference tasks with modest-dimension,
$d\sim\mathcal{O}(10)$.
Related papers
- Stein Variational Evolution Strategies [17.315583101484147]
Stein Variational Gradient Descent (SVGD) is a highly efficient method to sample from an unnormalized probability distribution.
Existing gradient-free versions of SVGD make use of simple Monte Carlo approximations or gradients from surrogate distributions, both with limitations.
We combine SVGD steps with evolution strategy (ES) updates to improve gradient-free Stein variational inference.
arXiv Detail & Related papers (2024-10-14T11:24:41Z) - Accelerating Convergence of Stein Variational Gradient Descent via Deep
Unfolding [5.584060970507506]
Stein variational gradient descent (SVGD) is a prominent particle-based variational inference method used for sampling a target distribution.
In this paper, we propose novel trainable algorithms that incorporate a deep-learning technique called deep unfolding,into SVGD.
arXiv Detail & Related papers (2024-02-23T06:24:57Z) - Learning Unnormalized Statistical Models via Compositional Optimization [73.30514599338407]
Noise-contrastive estimation(NCE) has been proposed by formulating the objective as the logistic loss of the real data and the artificial noise.
In this paper, we study it a direct approach for optimizing the negative log-likelihood of unnormalized models.
arXiv Detail & Related papers (2023-06-13T01:18:16Z) - Provably Fast Finite Particle Variants of SVGD via Virtual Particle
Stochastic Approximation [9.065034043031668]
Stein Variational Gradient Descent (SVGD) is a popular variational inference which simulates an interacting particle system to approximately sample from a target distribution.
We introduce the notion of virtual particles and develop novel approximations of population-limit dynamics in the space of probability measures.
We show that the $n$ particles output by VP-SVGD and GB-SVGD, run for $T$ steps with batch-size $K$, are as good as i.i.i.d samples from a distribution whose Kernel Stein Discrepancy to the target is at most $Oleft(tfrac
arXiv Detail & Related papers (2023-05-27T19:21:28Z) - Improved Convergence Rate of Stochastic Gradient Langevin Dynamics with
Variance Reduction and its Application to Optimization [50.83356836818667]
gradient Langevin Dynamics is one of the most fundamental algorithms to solve non-eps optimization problems.
In this paper, we show two variants of this kind, namely the Variance Reduced Langevin Dynamics and the Recursive Gradient Langevin Dynamics.
arXiv Detail & Related papers (2022-03-30T11:39:00Z) - Grassmann Stein Variational Gradient Descent [3.644031721554146]
Stein variational gradient descent (SVGD) is a deterministic particle inference algorithm that provides an efficient alternative to Markov chain Monte Carlo.
Recent developments have advocated projecting both the score function and the data onto real lines to sidestep this issue.
We propose Grassmann Stein variational gradient descent (GSVGD) as an alternative approach, which permits projections onto arbitrary dimensional subspaces.
arXiv Detail & Related papers (2022-02-07T15:36:03Z) - DyCo3D: Robust Instance Segmentation of 3D Point Clouds through Dynamic
Convolution [136.7261709896713]
We propose a data-driven approach that generates the appropriate convolution kernels to apply in response to the nature of the instances.
The proposed method achieves promising results on both ScanetNetV2 and S3DIS.
It also improves inference speed by more than 25% over the current state-of-the-art.
arXiv Detail & Related papers (2020-11-26T14:56:57Z) - Faster Convergence of Stochastic Gradient Langevin Dynamics for
Non-Log-Concave Sampling [110.88857917726276]
We provide a new convergence analysis of gradient Langevin dynamics (SGLD) for sampling from a class of distributions that can be non-log-concave.
At the core of our approach is a novel conductance analysis of SGLD using an auxiliary time-reversible Markov Chain.
arXiv Detail & Related papers (2020-10-19T15:23:18Z) - Kernel Stein Generative Modeling [68.03537693810972]
Gradient Langevin Dynamics (SGLD) demonstrates impressive results with energy-based models on high-dimensional and complex data distributions.
Stein Variational Gradient Descent (SVGD) is a deterministic sampling algorithm that iteratively transports a set of particles to approximate a given distribution.
We propose noise conditional kernel SVGD (NCK-SVGD), that works in tandem with the recently introduced Noise Conditional Score Network estimator.
arXiv Detail & Related papers (2020-07-06T21:26:04Z) - Stein Variational Inference for Discrete Distributions [70.19352762933259]
We propose a simple yet general framework that transforms discrete distributions to equivalent piecewise continuous distributions.
Our method outperforms traditional algorithms such as Gibbs sampling and discontinuous Hamiltonian Monte Carlo.
We demonstrate that our method provides a promising tool for learning ensembles of binarized neural network (BNN)
In addition, such transform can be straightforwardly employed in gradient-free kernelized Stein discrepancy to perform goodness-of-fit (GOF) test on discrete distributions.
arXiv Detail & Related papers (2020-03-01T22:45:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.