SwISS: A Scalable Markov chain Monte Carlo Divide-and-Conquer Strategy
- URL: http://arxiv.org/abs/2208.04080v1
- Date: Mon, 8 Aug 2022 12:02:18 GMT
- Title: SwISS: A Scalable Markov chain Monte Carlo Divide-and-Conquer Strategy
- Authors: Callum Vyner, Christopher Nemeth, Chris Sherlock
- Abstract summary: Divide-and-conquer strategies for Monte Carlo algorithms are an increasingly popular approach to making Bayesian inference scalable to large data sets.
We propose SwISS: Sub-posteriors with Inflation, Scaling and Shifting; a new approach for recombining the sub-posterior samples.
We prove that our transformation is optimal across a natural set of affine transformations and illustrate the efficacy of SwISS against competing algorithms on synthetic and real-world data sets.
- Score: 1.6114012813668934
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Divide-and-conquer strategies for Monte Carlo algorithms are an increasingly
popular approach to making Bayesian inference scalable to large data sets. In
its simplest form, the data are partitioned across multiple computing cores and
a separate Markov chain Monte Carlo algorithm on each core targets the
associated partial posterior distribution, which we refer to as a
sub-posterior, that is the posterior given only the data from the segment of
the partition associated with that core. Divide-and-conquer techniques reduce
computational, memory and disk bottle-necks, but make it difficult to recombine
the sub-posterior samples. We propose SwISS: Sub-posteriors with Inflation,
Scaling and Shifting; a new approach for recombining the sub-posterior samples
which is simple to apply, scales to high-dimensional parameter spaces and
accurately approximates the original posterior distribution through affine
transformations of the sub-posterior samples. We prove that our transformation
is asymptotically optimal across a natural set of affine transformations and
illustrate the efficacy of SwISS against competing algorithms on synthetic and
real-world data sets.
Related papers
- Variational Learning of Gaussian Process Latent Variable Models through Stochastic Gradient Annealed Importance Sampling [22.256068524699472]
In this work, we propose an Annealed Importance Sampling (AIS) approach to address these issues.
We combine the strengths of Sequential Monte Carlo samplers and VI to explore a wider range of posterior distributions and gradually approach the target distribution.
Experimental results on both toy and image datasets demonstrate that our method outperforms state-of-the-art methods in terms of tighter variational bounds, higher log-likelihoods, and more robust convergence.
arXiv Detail & Related papers (2024-08-13T08:09:05Z) - An Efficient Algorithm for Clustered Multi-Task Compressive Sensing [60.70532293880842]
Clustered multi-task compressive sensing is a hierarchical model that solves multiple compressive sensing tasks.
The existing inference algorithm for this model is computationally expensive and does not scale well in high dimensions.
We propose a new algorithm that substantially accelerates model inference by avoiding the need to explicitly compute these covariance matrices.
arXiv Detail & Related papers (2023-09-30T15:57:14Z) - Langevin Monte Carlo for Contextual Bandits [72.00524614312002]
Langevin Monte Carlo Thompson Sampling (LMC-TS) is proposed to directly sample from the posterior distribution in contextual bandits.
We prove that the proposed algorithm achieves the same sublinear regret bound as the best Thompson sampling algorithms for a special case of contextual bandits.
arXiv Detail & Related papers (2022-06-22T17:58:23Z) - Change-point Detection and Segmentation of Discrete Data using Bayesian
Context Trees [7.090165638014331]
Building on the recently introduced Bayesian Context Trees (BCT) framework, the distributions of different segments in a discrete time series are described as variable-memory Markov chains.
Inference for the presence and location of change-points is then performed via Markov chain Monte Carlo sampling.
Results on both simulated and real-world data indicate that the proposed methodology performs better than or as well as state-of-the-art techniques.
arXiv Detail & Related papers (2022-03-08T19:03:21Z) - Improving the Sample-Complexity of Deep Classification Networks with
Invariant Integration [77.99182201815763]
Leveraging prior knowledge on intraclass variance due to transformations is a powerful method to improve the sample complexity of deep neural networks.
We propose a novel monomial selection algorithm based on pruning methods to allow an application to more complex problems.
We demonstrate the improved sample complexity on the Rotated-MNIST, SVHN and CIFAR-10 datasets.
arXiv Detail & Related papers (2022-02-08T16:16:11Z) - Inferring Manifolds From Noisy Data Using Gaussian Processes [17.166283428199634]
Most existing manifold learning algorithms replace the original data with lower dimensional coordinates.
This article proposes a new methodology for addressing these problems, allowing the estimated manifold between fitted data points.
arXiv Detail & Related papers (2021-10-14T15:50:38Z) - Sampling in Combinatorial Spaces with SurVAE Flow Augmented MCMC [83.48593305367523]
Hybrid Monte Carlo is a powerful Markov Chain Monte Carlo method for sampling from complex continuous distributions.
We introduce a new approach based on augmenting Monte Carlo methods with SurVAE Flows to sample from discrete distributions.
We demonstrate the efficacy of our algorithm on a range of examples from statistics, computational physics and machine learning, and observe improvements compared to alternative algorithms.
arXiv Detail & Related papers (2021-02-04T02:21:08Z) - Kernel learning approaches for summarising and combining posterior
similarity matrices [68.8204255655161]
We build upon the notion of the posterior similarity matrix (PSM) in order to suggest new approaches for summarising the output of MCMC algorithms for Bayesian clustering models.
A key contribution of our work is the observation that PSMs are positive semi-definite, and hence can be used to define probabilistically-motivated kernel matrices.
arXiv Detail & Related papers (2020-09-27T14:16:14Z) - Multilevel Gibbs Sampling for Bayesian Regression [6.2997667081978825]
The level hierarchy of data matrices is created by clustering the features and/or samples of data matrices.
The use of correlated samples is investigated for variance reduction to improve the convergence of the Markov Chain.
Speed-up is achieved for almost all of them without significant loss in predictive performance.
arXiv Detail & Related papers (2020-09-25T11:18:17Z) - Model Fusion with Kullback--Leibler Divergence [58.20269014662046]
We propose a method to fuse posterior distributions learned from heterogeneous datasets.
Our algorithm relies on a mean field assumption for both the fused model and the individual dataset posteriors.
arXiv Detail & Related papers (2020-07-13T03:27:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.