Reversible Jump PDMP Samplers for Variable Selection
- URL: http://arxiv.org/abs/2010.11771v1
- Date: Thu, 22 Oct 2020 14:46:33 GMT
- Title: Reversible Jump PDMP Samplers for Variable Selection
- Authors: Augustin Chevallier, Paul Fearnhead, Matthew Sutton
- Abstract summary: A new class of Monte Carlo algorithms based on piecewise deterministic Markov processes (PDMPs) have recently shown great promise.
PDMP samplers can only sample from posterior densities that are differentiable almost everywhere.
We show how to develop reversible jump PDMP samplers that can jointly explore the discrete space of models and the continuous space of parameters.
- Score: 1.5469452301122175
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: A new class of Markov chain Monte Carlo (MCMC) algorithms, based on
simulating piecewise deterministic Markov processes (PDMPs), have recently
shown great promise: they are non-reversible, can mix better than standard MCMC
algorithms, and can use subsampling ideas to speed up computation in big data
scenarios. However, current PDMP samplers can only sample from posterior
densities that are differentiable almost everywhere, which precludes their use
for model choice. Motivated by variable selection problems, we show how to
develop reversible jump PDMP samplers that can jointly explore the discrete
space of models and the continuous space of parameters. Our framework is
general: it takes any existing PDMP sampler, and adds two types of
trans-dimensional moves that allow for the addition or removal of a variable
from the model. We show how the rates of these trans-dimensional moves can be
calculated so that the sampler has the correct invariant distribution.
Simulations show that the new samplers can mix better than standard MCMC
algorithms. Our empirical results show they are also more efficient than
gradient-based samplers that avoid model choice through use of continuous
spike-and-slab priors which replace a point mass at zero for each parameter
with a density concentrated around zero.
Related papers
- Stochastic Sampling from Deterministic Flow Models [8.849981177332594]
We present a method to turn flow models into a family of differential equations (SDEs) that have the same marginal distributions.
We empirically demonstrate advantages of our method on a toy Gaussian setup and on the large scale ImageNet generation task.
arXiv Detail & Related papers (2024-10-03T05:18:28Z) - Stochastic Gradient Piecewise Deterministic Monte Carlo Samplers [3.487370856323828]
Recent work has suggested using Monte Carlo methods based on piecewise deterministic Markov processes (PDMPs) to sample from target distributions of interest.
We propose approximate simulation of PDMPs with sub-sampling for scalable sampling from posterior distributions.
We show these methods are easy to implement, present results on their approximation error and demonstrate numerically that this class of algorithms has similar efficiency to gradient Langevin dynamics.
arXiv Detail & Related papers (2024-06-27T09:59:28Z) - A Block Metropolis-Hastings Sampler for Controllable Energy-based Text
Generation [78.81021361497311]
We develop a novel Metropolis-Hastings (MH) sampler that proposes re-writes of the entire sequence in each step via iterative prompting of a large language model.
Our new sampler allows for more efficient and accurate sampling from a target distribution and (b) allows generation length to be determined through the sampling procedure rather than fixed in advance.
arXiv Detail & Related papers (2023-12-07T18:30:15Z) - AdjointDPM: Adjoint Sensitivity Method for Gradient Backpropagation of Diffusion Probabilistic Models [103.41269503488546]
Existing customization methods require access to multiple reference examples to align pre-trained diffusion probabilistic models with user-provided concepts.
This paper aims to address the challenge of DPM customization when the only available supervision is a differentiable metric defined on the generated contents.
We propose a novel method AdjointDPM, which first generates new samples from diffusion models by solving the corresponding probability-flow ODEs.
It then uses the adjoint sensitivity method to backpropagate the gradients of the loss to the models' parameters.
arXiv Detail & Related papers (2023-07-20T09:06:21Z) - On Calibrating Diffusion Probabilistic Models [78.75538484265292]
diffusion probabilistic models (DPMs) have achieved promising results in diverse generative tasks.
We propose a simple way for calibrating an arbitrary pretrained DPM, with which the score matching loss can be reduced and the lower bounds of model likelihood can be increased.
Our calibration method is performed only once and the resulting models can be used repeatedly for sampling.
arXiv Detail & Related papers (2023-02-21T14:14:40Z) - Learning Sampling Distributions for Model Predictive Control [36.82905770866734]
Sampling-based approaches to Model Predictive Control (MPC) have become a cornerstone of contemporary approaches to MPC.
We propose to carry out all operations in the latent space, allowing us to take full advantage of the learned distribution.
Specifically, we frame the learning problem as bi-level optimization and show how to train the controller with backpropagation-through-time.
arXiv Detail & Related papers (2022-12-05T20:35:36Z) - A new perspective on probabilistic image modeling [92.89846887298852]
We present a new probabilistic approach for image modeling capable of density estimation, sampling and tractable inference.
DCGMMs can be trained end-to-end by SGD from random initial conditions, much like CNNs.
We show that DCGMMs compare favorably to several recent PC and SPN models in terms of inference, classification and sampling.
arXiv Detail & Related papers (2022-03-21T14:53:57Z) - Sampling Approximately Low-Rank Ising Models: MCMC meets Variational
Methods [35.24886589614034]
We consider quadratic definite Ising models on the hypercube with a general interaction $J$.
Our general result implies the first time sampling algorithms for low-rank Ising models.
arXiv Detail & Related papers (2022-02-17T21:43:50Z) - Efficient Learning of the Parameters of Non-Linear Models using
Differentiable Resampling in Particle Filters [1.9499120576896227]
It has been widely documented that the sampling and resampling steps in particle filters be differentiated.
We consider two state-space models and show that NUTS improves the mixing of the Markov chain and can produce more accurate results in less computational time.
arXiv Detail & Related papers (2021-11-02T08:03:09Z) - Sampling from Arbitrary Functions via PSD Models [55.41644538483948]
We take a two-step approach by first modeling the probability distribution and then sampling from that model.
We show that these models can approximate a large class of densities concisely using few evaluations, and present a simple algorithm to effectively sample from these models.
arXiv Detail & Related papers (2021-10-20T12:25:22Z) - Oops I Took A Gradient: Scalable Sampling for Discrete Distributions [53.3142984019796]
We show that this approach outperforms generic samplers in a number of difficult settings.
We also demonstrate the use of our improved sampler for training deep energy-based models on high dimensional discrete data.
arXiv Detail & Related papers (2021-02-08T20:08:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.