Toward Unlimited Self-Learning MCMC with Parallel Adaptive Annealing
- URL: http://arxiv.org/abs/2211.14024v2
- Date: Wed, 20 Sep 2023 01:17:05 GMT
- Title: Toward Unlimited Self-Learning MCMC with Parallel Adaptive Annealing
- Authors: Yuma Ichikawa, Akira Nakagawa, Hiromoto Masayuki, Yuhei Umeda
- Abstract summary: Self-learning Monte Carlo (SLMC) methods are recently proposed to accelerate Markov chain Monte Carlo (MCMC) methods using a machine learning model.
With latent generative models, SLMC methods realize efficient Monte Carlo updates with less autocorrelation.
However, SLMC methods are difficult to directly apply to multimodal distributions for which training data are difficult to obtain.
- Score: 4.156535226615696
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Self-learning Monte Carlo (SLMC) methods are recently proposed to accelerate
Markov chain Monte Carlo (MCMC) methods using a machine learning model. With
latent generative models, SLMC methods realize efficient Monte Carlo updates
with less autocorrelation. However, SLMC methods are difficult to directly
apply to multimodal distributions for which training data are difficult to
obtain. To solve the limitation, we propose parallel adaptive annealing, which
makes SLMC methods directly apply to multimodal distributions with a gradually
trained proposal while annealing target distribution. Parallel adaptive
annealing is based on (i) sequential learning with annealing to inherit and
update the model parameters, (ii) adaptive annealing to automatically detect
under-learning, and (iii) parallel annealing to mitigate mode collapse of
proposal models. We also propose VAE-SLMC method which utilizes a variational
autoencoder (VAE) as a proposal of SLMC to make efficient parallel proposals
independent of any previous state using recently clarified quantitative
properties of VAE. Experiments validate that our method can proficiently obtain
accurate samples from multiple multimodal toy distributions and practical
multimodal posterior distributions, which is difficult to achieve with the
existing SLMC methods.
Related papers
- Recursive Learning of Asymptotic Variational Objectives [49.69399307452126]
General state-space models (SSMs) are widely used in statistical machine learning and are among the most classical generative models for sequential time-series data.
Online sequential IWAE (OSIWAE) allows for online learning of both model parameters and a Markovian recognition model for inferring latent states.
This approach is more theoretically well-founded than recently proposed online variational SMC methods.
arXiv Detail & Related papers (2024-11-04T16:12:37Z) - AutoStep: Locally adaptive involutive MCMC [51.186543293659376]
AutoStep MCMC selects an appropriate step size at each iteration adapted to the local geometry of the target distribution.
We show that AutoStep MCMC is competitive with state-of-the-art methods in terms of effective sample size per unit cost.
arXiv Detail & Related papers (2024-10-24T17:17:11Z) - Convergence Acceleration of Markov Chain Monte Carlo-based Gradient
Descent by Deep Unfolding [5.584060970507506]
This study proposes a trainable sampling-based solver for optimization problems (COPs) using a deep-learning technique called deep unfolding.
The proposed solver is based on the Ohzeki method that combines Markov-chain Monte-Carlo (MCMC) and gradient descent.
The numerical results for a few COPs demonstrated that the proposed solver significantly accelerated the convergence speed compared with the original Ohzeki method.
arXiv Detail & Related papers (2024-02-21T08:21:48Z) - Online Variational Sequential Monte Carlo [49.97673761305336]
We build upon the variational sequential Monte Carlo (VSMC) method, which provides computationally efficient and accurate model parameter estimation and Bayesian latent-state inference.
Online VSMC is capable of performing efficiently, entirely on-the-fly, both parameter estimation and particle proposal adaptation.
arXiv Detail & Related papers (2023-12-19T21:45:38Z) - Reverse Diffusion Monte Carlo [19.35592726471155]
We propose a novel Monte Carlo sampling algorithm called reverse diffusion Monte Carlo (rdMC)
rdMC is distinct from the Markov chain Monte Carlo (MCMC) methods.
arXiv Detail & Related papers (2023-07-05T05:42:03Z) - Online Probabilistic Model Identification using Adaptive Recursive MCMC [8.465242072268019]
We suggest the Adaptive Recursive Markov Chain Monte Carlo (ARMCMC) method.
It eliminates the shortcomings of conventional online techniques while computing the entire probability density function of model parameters.
We demonstrate our approach using parameter estimation in a soft bending actuator and the Hunt-Crossley dynamic model.
arXiv Detail & Related papers (2022-10-23T02:06:48Z) - Sampling in Combinatorial Spaces with SurVAE Flow Augmented MCMC [83.48593305367523]
Hybrid Monte Carlo is a powerful Markov Chain Monte Carlo method for sampling from complex continuous distributions.
We introduce a new approach based on augmenting Monte Carlo methods with SurVAE Flows to sample from discrete distributions.
We demonstrate the efficacy of our algorithm on a range of examples from statistics, computational physics and machine learning, and observe improvements compared to alternative algorithms.
arXiv Detail & Related papers (2021-02-04T02:21:08Z) - No MCMC for me: Amortized sampling for fast and stable training of
energy-based models [62.1234885852552]
Energy-Based Models (EBMs) present a flexible and appealing way to represent uncertainty.
We present a simple method for training EBMs at scale using an entropy-regularized generator to amortize the MCMC sampling.
Next, we apply our estimator to the recently proposed Joint Energy Model (JEM), where we match the original performance with faster and stable training.
arXiv Detail & Related papers (2020-10-08T19:17:20Z) - Non-convex Learning via Replica Exchange Stochastic Gradient MCMC [25.47669573608621]
We propose an adaptive replica exchange SGMCMC (reSGMCMC) to automatically correct the bias and study the corresponding properties.
Empirically, we test the algorithm through extensive experiments on various setups and obtain the results.
arXiv Detail & Related papers (2020-08-12T15:02:59Z) - Adaptive Correlated Monte Carlo for Contextual Categorical Sequence
Generation [77.7420231319632]
We adapt contextual generation of categorical sequences to a policy gradient estimator, which evaluates a set of correlated Monte Carlo (MC) rollouts for variance control.
We also demonstrate the use of correlated MC rollouts for binary-tree softmax models, which reduce the high generation cost in large vocabulary scenarios.
arXiv Detail & Related papers (2019-12-31T03:01:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.