Parallel MCMC Without Embarrassing Failures
- URL: http://arxiv.org/abs/2202.11154v1
- Date: Tue, 22 Feb 2022 20:17:46 GMT
- Title: Parallel MCMC Without Embarrassing Failures
- Authors: Daniel Augusto de Souza, Diego Mesquita, Samuel Kaski, Luigi Acerbi
- Abstract summary: MCMC is run in parallel on (sub)posteriors defined on data partitions.
While efficient, this framework is very sensitive to the quality of subposterior sampling.
We propose a novel combination strategy to mitigate this issue.
- Score: 19.429985676081618
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Embarrassingly parallel Markov Chain Monte Carlo (MCMC) exploits parallel
computing to scale Bayesian inference to large datasets by using a two-step
approach. First, MCMC is run in parallel on (sub)posteriors defined on data
partitions. Then, a server combines local results. While efficient, this
framework is very sensitive to the quality of subposterior sampling. Common
sampling problems such as missing modes or misrepresentation of low-density
regions are amplified -- instead of being corrected -- in the combination
phase, leading to catastrophic failures. In this work, we propose a novel
combination strategy to mitigate this issue. Our strategy, Parallel Active
Inference (PAI), leverages Gaussian Process (GP) surrogate modeling and active
learning. After fitting GPs to subposteriors, PAI (i) shares information
between GP surrogates to cover missing modes; and (ii) uses active sampling to
individually refine subposterior approximations. We validate PAI in challenging
benchmarks, including heavy-tailed and multi-modal posteriors and a real-world
application to computational neuroscience. Empirical results show that PAI
succeeds where previous methods catastrophically fail, with a small
communication overhead.
Related papers
- A Bayesian Approach to Data Point Selection [24.98069363998565]
Data point selection (DPS) is becoming a critical topic in deep learning.
Existing approaches to DPS are predominantly based on a bi-level optimisation (BLO) formulation.
We propose a novel Bayesian approach to DPS.
arXiv Detail & Related papers (2024-11-06T09:04:13Z) - Diffusion Generative Modelling for Divide-and-Conquer MCMC [0.0]
Divide-and-conquer MCMC is a strategy for parallelising Markov Chain Monte Carlo sampling.
We propose using diffusion generative modelling to fit density approximations to the subposterior distributions.
arXiv Detail & Related papers (2024-06-17T15:48:46Z) - Online Variational Sequential Monte Carlo [49.97673761305336]
We build upon the variational sequential Monte Carlo (VSMC) method, which provides computationally efficient and accurate model parameter estimation and Bayesian latent-state inference.
Online VSMC is capable of performing efficiently, entirely on-the-fly, both parameter estimation and particle proposal adaptation.
arXiv Detail & Related papers (2023-12-19T21:45:38Z) - Overlap-guided Gaussian Mixture Models for Point Cloud Registration [61.250516170418784]
Probabilistic 3D point cloud registration methods have shown competitive performance in overcoming noise, outliers, and density variations.
This paper proposes a novel overlap-guided probabilistic registration approach that computes the optimal transformation from matched Gaussian Mixture Model (GMM) parameters.
arXiv Detail & Related papers (2022-10-17T08:02:33Z) - Compressed Monte Carlo with application in particle filtering [11.84836209560411]
We introduce the theory and practice of a Compressed MC (C-MC) scheme to compress the statistical information contained in a set of random samples.
C-MC is useful within particle filtering and adaptive IS algorithms, as shown by three novel schemes introduced in this work.
arXiv Detail & Related papers (2021-07-18T14:32:04Z) - Semi-Supervised Learning with Variational Bayesian Inference and Maximum
Uncertainty Regularization [62.21716612888669]
We propose two generic methods for improving semi-supervised learning (SSL)
The first integrates weight perturbation (WP) into existing "consistency regularization" (CR) based methods.
The second method proposes a novel consistency loss called "maximum uncertainty regularization" (MUR)
arXiv Detail & Related papers (2020-12-03T09:49:35Z) - PaDiM: a Patch Distribution Modeling Framework for Anomaly Detection and
Localization [64.39761523935613]
We present a new framework for Patch Distribution Modeling, PaDiM, to concurrently detect and localize anomalies in images.
PaDiM makes use of a pretrained convolutional neural network (CNN) for patch embedding.
It also exploits correlations between the different semantic levels of CNN to better localize anomalies.
arXiv Detail & Related papers (2020-11-17T17:29:18Z) - An adaptive Hessian approximated stochastic gradient MCMC method [12.93317525451798]
We present an adaptive Hessian approximated gradient MCMC method to incorporate local geometric information while sampling from the posterior.
We adopt a magnitude-based weight pruning method to enforce the sparsity of the network.
arXiv Detail & Related papers (2020-10-03T16:22:15Z) - Coded Stochastic ADMM for Decentralized Consensus Optimization with Edge
Computing [113.52575069030192]
Big data, including applications with high security requirements, are often collected and stored on multiple heterogeneous devices, such as mobile devices, drones and vehicles.
Due to the limitations of communication costs and security requirements, it is of paramount importance to extract information in a decentralized manner instead of aggregating data to a fusion center.
We consider the problem of learning model parameters in a multi-agent system with data locally processed via distributed edge nodes.
A class of mini-batch alternating direction method of multipliers (ADMM) algorithms is explored to develop the distributed learning model.
arXiv Detail & Related papers (2020-10-02T10:41:59Z) - Revisiting LSTM Networks for Semi-Supervised Text Classification via
Mixed Objective Function [106.69643619725652]
We develop a training strategy that allows even a simple BiLSTM model, when trained with cross-entropy loss, to achieve competitive results.
We report state-of-the-art results for text classification task on several benchmark datasets.
arXiv Detail & Related papers (2020-09-08T21:55:22Z) - Stacking for Non-mixing Bayesian Computations: The Curse and Blessing of
Multimodal Posteriors [8.11978827493967]
We propose an approach using parallel runs of MCMC, variational, or mode-based inference to hit as many modes as possible.
We present theoretical consistency with an example where the stacked inference process approximates the true data.
We demonstrate practical implementation in several model families.
arXiv Detail & Related papers (2020-06-22T15:26:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.