Decentralized Stochastic Gradient Langevin Dynamics and Hamiltonian
Monte Carlo
- URL: http://arxiv.org/abs/2007.00590v4
- Date: Thu, 26 Aug 2021 19:46:03 GMT
- Title: Decentralized Stochastic Gradient Langevin Dynamics and Hamiltonian
Monte Carlo
- Authors: Mert G\"urb\"uzbalaban, Xuefeng Gao, Yuanhan Hu, Lingjiong Zhu
- Abstract summary: Decentralized SGLD (DE-SGLD) and Decentralized SGHMC (DE-SGHMC) are algorithms for scaleable Bayesian inference in the decentralized setting for large datasets.
We show that when the posterior distribution is strongly log-concave and smooth, the iterates of these algorithms converge linearly to a neighborhood of the target distribution in the 2-Wasserstein distance.
- Score: 8.94392435424862
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Stochastic gradient Langevin dynamics (SGLD) and stochastic gradient
Hamiltonian Monte Carlo (SGHMC) are two popular Markov Chain Monte Carlo (MCMC)
algorithms for Bayesian inference that can scale to large datasets, allowing to
sample from the posterior distribution of the parameters of a statistical model
given the input data and the prior distribution over the model parameters.
However, these algorithms do not apply to the decentralized learning setting,
when a network of agents are working collaboratively to learn the parameters of
a statistical model without sharing their individual data due to privacy
reasons or communication constraints. We study two algorithms: Decentralized
SGLD (DE-SGLD) and Decentralized SGHMC (DE-SGHMC) which are adaptations of SGLD
and SGHMC methods that allow scaleable Bayesian inference in the decentralized
setting for large datasets. We show that when the posterior distribution is
strongly log-concave and smooth, the iterates of these algorithms converge
linearly to a neighborhood of the target distribution in the 2-Wasserstein
distance if their parameters are selected appropriately. We illustrate the
efficiency of our algorithms on decentralized Bayesian linear regression and
Bayesian logistic regression problems.
Related papers
- Latent Semantic Consensus For Deterministic Geometric Model Fitting [109.44565542031384]
We propose an effective method called Latent Semantic Consensus (LSC)
LSC formulates the model fitting problem into two latent semantic spaces based on data points and model hypotheses.
LSC is able to provide consistent and reliable solutions within only a few milliseconds for general multi-structural model fitting.
arXiv Detail & Related papers (2024-03-11T05:35:38Z) - Online Variational Sequential Monte Carlo [49.97673761305336]
We build upon the variational sequential Monte Carlo (VSMC) method, which provides computationally efficient and accurate model parameter estimation and Bayesian latent-state inference.
Online VSMC is capable of performing efficiently, entirely on-the-fly, both parameter estimation and particle proposal adaptation.
arXiv Detail & Related papers (2023-12-19T21:45:38Z) - Stability and Generalization of the Decentralized Stochastic Gradient
Descent Ascent Algorithm [80.94861441583275]
We investigate the complexity of the generalization bound of the decentralized gradient descent (D-SGDA) algorithm.
Our results analyze the impact of different top factors on the generalization of D-SGDA.
We also balance it with the generalization to obtain the optimal convex-concave setting.
arXiv Detail & Related papers (2023-10-31T11:27:01Z) - Spike-and-Slab Generalized Additive Models and Scalable Algorithms for
High-Dimensional Data [0.0]
We propose hierarchical generalized additive models (GAMs) to accommodate high-dimensional data.
We consider the smoothing penalty for proper shrinkage of curve and separation of smoothing function linear and nonlinear spaces.
Two and deterministic algorithms, EM-Coordinate Descent and EM-Iterative Weighted Least Squares, are developed for different utilities.
arXiv Detail & Related papers (2021-10-27T14:11:13Z) - What Are Bayesian Neural Network Posteriors Really Like? [63.950151520585024]
We show that Hamiltonian Monte Carlo can achieve significant performance gains over standard and deep ensembles.
We also show that deep distributions are similarly close to HMC as standard SGLD, and closer than standard variational inference.
arXiv Detail & Related papers (2021-04-29T15:38:46Z) - Bayesian multiscale deep generative model for the solution of
high-dimensional inverse problems [0.0]
A novel multiscale Bayesian inference approach is introduced based on deep probabilistic generative models.
The method allows high-dimensional parameter estimation while exhibiting stability, efficiency and accuracy.
arXiv Detail & Related papers (2021-02-04T11:47:21Z) - A Contour Stochastic Gradient Langevin Dynamics Algorithm for
Simulations of Multi-modal Distributions [17.14287157979558]
We propose an adaptively weighted gradient Langevin dynamics (SGLD) for learning in big data statistics.
The proposed algorithm is tested on benchmark datasets including CIFAR100.
arXiv Detail & Related papers (2020-10-19T19:20:47Z) - An adaptive Hessian approximated stochastic gradient MCMC method [12.93317525451798]
We present an adaptive Hessian approximated gradient MCMC method to incorporate local geometric information while sampling from the posterior.
We adopt a magnitude-based weight pruning method to enforce the sparsity of the network.
arXiv Detail & Related papers (2020-10-03T16:22:15Z) - Coded Stochastic ADMM for Decentralized Consensus Optimization with Edge
Computing [113.52575069030192]
Big data, including applications with high security requirements, are often collected and stored on multiple heterogeneous devices, such as mobile devices, drones and vehicles.
Due to the limitations of communication costs and security requirements, it is of paramount importance to extract information in a decentralized manner instead of aggregating data to a fusion center.
We consider the problem of learning model parameters in a multi-agent system with data locally processed via distributed edge nodes.
A class of mini-batch alternating direction method of multipliers (ADMM) algorithms is explored to develop the distributed learning model.
arXiv Detail & Related papers (2020-10-02T10:41:59Z) - Decentralized MCTS via Learned Teammate Models [89.24858306636816]
We present a trainable online decentralized planning algorithm based on decentralized Monte Carlo Tree Search.
We show that deep learning and convolutional neural networks can be employed to produce accurate policy approximators.
arXiv Detail & Related papers (2020-03-19T13:10:20Z) - DDKSP: A Data-Driven Stochastic Programming Framework for Car-Sharing
Relocation Problem [17.440172040605354]
We investigate the car-sharing relocation problem (CSRP) under uncertain demands.
In order to overcome the problem, an innovative framework called Data-Driven Kernel Programming (DDKSP) is proposed.
The proposed framework outperforms the pure parametric approaches with 3.72%, 4.58% and 11% in terms of overall profits.
arXiv Detail & Related papers (2020-01-20T19:04:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.