MetFlow: A New Efficient Method for Bridging the Gap between Markov
Chain Monte Carlo and Variational Inference
- URL: http://arxiv.org/abs/2002.12253v1
- Date: Thu, 27 Feb 2020 16:50:30 GMT
- Title: MetFlow: A New Efficient Method for Bridging the Gap between Markov
Chain Monte Carlo and Variational Inference
- Authors: Achille Thin, Nikita Kotelevskii, Jean-Stanislas Denain, Leo
Grinsztajn, Alain Durmus, Maxim Panov and Eric Moulines
- Abstract summary: We propose a new computationally efficient method to combine Variational Inference (VI) with Markov Chain Monte Carlo (MCMC)
This approach can be used with generic MCMC kernels, but is especially well suited to textitMetFlow, a novel family of MCMC algorithms we introduce.
- Score: 20.312106392307406
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this contribution, we propose a new computationally efficient method to
combine Variational Inference (VI) with Markov Chain Monte Carlo (MCMC). This
approach can be used with generic MCMC kernels, but is especially well suited
to \textit{MetFlow}, a novel family of MCMC algorithms we introduce, in which
proposals are obtained using Normalizing Flows. The marginal distribution
produced by such MCMC algorithms is a mixture of flow-based distributions, thus
drastically increasing the expressivity of the variational family. Unlike
previous methods following this direction, our approach is amenable to the
reparametrization trick and does not rely on computationally expensive reverse
kernels. Extensive numerical experiments show clear computational and
performance improvements over state-of-the-art methods.
Related papers
- Markovian Flow Matching: Accelerating MCMC with Continuous Normalizing Flows [2.2530496464901106]
Continuous normalizing flows (CNFs) learn the probability path between a reference distribution and a target distribution by modeling the vector field generating said path using neural networks.
Recently, Lipman et al. (2022) introduced a simple and inexpensive method for training CNFs in generative modeling, termed flow matching (FM)
In this paper, we repurpose this method for probabilistic inference by incorporating Markovian sampling methods in evaluating the FM objective, and using the learned CNF to improve Monte Carlo sampling.
arXiv Detail & Related papers (2024-05-23T10:08:19Z) - Fast Semisupervised Unmixing Using Nonconvex Optimization [80.11512905623417]
We introduce a novel convex convex model for semi/library-based unmixing.
We demonstrate the efficacy of Alternating Methods of sparse unsupervised unmixing.
arXiv Detail & Related papers (2024-01-23T10:07:41Z) - Online Variational Sequential Monte Carlo [49.97673761305336]
We build upon the variational sequential Monte Carlo (VSMC) method, which provides computationally efficient and accurate model parameter estimation and Bayesian latent-state inference.
Online VSMC is capable of performing efficiently, entirely on-the-fly, both parameter estimation and particle proposal adaptation.
arXiv Detail & Related papers (2023-12-19T21:45:38Z) - Reverse Diffusion Monte Carlo [19.35592726471155]
We propose a novel Monte Carlo sampling algorithm called reverse diffusion Monte Carlo (rdMC)
rdMC is distinct from the Markov chain Monte Carlo (MCMC) methods.
arXiv Detail & Related papers (2023-07-05T05:42:03Z) - Fast Doubly-Adaptive MCMC to Estimate the Gibbs Partition Function with
Weak Mixing Time Bounds [7.428782604099876]
A major obstacle to practical applications of Gibbs distributions is the need to estimate their partition functions.
We present a novel method for reducing the computational complexity of rigorously estimating the partition functions.
arXiv Detail & Related papers (2021-11-14T15:42:02Z) - Scalable Variational Gaussian Processes via Harmonic Kernel
Decomposition [54.07797071198249]
We introduce a new scalable variational Gaussian process approximation which provides a high fidelity approximation while retaining general applicability.
We demonstrate that, on a range of regression and classification problems, our approach can exploit input space symmetries such as translations and reflections.
Notably, our approach achieves state-of-the-art results on CIFAR-10 among pure GP models.
arXiv Detail & Related papers (2021-06-10T18:17:57Z) - Sampling in Combinatorial Spaces with SurVAE Flow Augmented MCMC [83.48593305367523]
Hybrid Monte Carlo is a powerful Markov Chain Monte Carlo method for sampling from complex continuous distributions.
We introduce a new approach based on augmenting Monte Carlo methods with SurVAE Flows to sample from discrete distributions.
We demonstrate the efficacy of our algorithm on a range of examples from statistics, computational physics and machine learning, and observe improvements compared to alternative algorithms.
arXiv Detail & Related papers (2021-02-04T02:21:08Z) - Accelerating MCMC algorithms through Bayesian Deep Networks [7.054093620465401]
Markov Chain Monte Carlo (MCMC) algorithms are commonly used for their versatility in sampling from complicated probability distributions.
As the dimension of the distribution gets larger, the computational costs for a satisfactory exploration of the sampling space become challenging.
We show an alternative way of performing adaptive MCMC, by using the outcome of Bayesian Neural Networks as the initial proposal for the Markov Chain.
arXiv Detail & Related papers (2020-11-29T04:29:00Z) - MCMC-Interactive Variational Inference [56.58416764959414]
We propose MCMC-interactive variational inference (MIVI) to estimate the posterior in a time constrained manner.
MIVI takes advantage of the complementary properties of variational inference and MCMC to encourage mutual improvement.
Experiments show that MIVI not only accurately approximates the posteriors but also facilitates designs of gradient MCMC and Gibbs sampling transitions.
arXiv Detail & Related papers (2020-10-02T17:43:20Z) - Involutive MCMC: a Unifying Framework [64.46316409766764]
We describe a wide range of MCMC algorithms in terms of iMCMC.
We formulate a number of "tricks" which one can use as design principles for developing new MCMC algorithms.
We demonstrate the latter with two examples where we transform known reversible MCMC algorithms into more efficient irreversible ones.
arXiv Detail & Related papers (2020-06-30T10:21:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.