Nonparametric Hamiltonian Monte Carlo
- URL: http://arxiv.org/abs/2106.10238v1
- Date: Fri, 18 Jun 2021 17:03:05 GMT
- Title: Nonparametric Hamiltonian Monte Carlo
- Authors: Carol Mak, Fabian Zaiser, Luke Ong
- Abstract summary: This paper introduces the Non Hamiltonian Monte Carlo (NP-HMC) algorithm which generalises HMC to nonparametric models.
We provide a correctness proof of NP-HMC, and empirically demonstrate significant performance improvements over existing approaches on several nonparametric examples.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Probabilistic programming uses programs to express generative models whose
posterior probability is then computed by built-in inference engines. A
challenging goal is to develop general purpose inference algorithms that work
out-of-the-box for arbitrary programs in a universal probabilistic programming
language (PPL). The densities defined by such programs, which may use
stochastic branching and recursion, are (in general) nonparametric, in the
sense that they correspond to models on an infinite-dimensional parameter
space. However standard inference algorithms, such as the Hamiltonian Monte
Carlo (HMC) algorithm, target distributions with a fixed number of parameters.
This paper introduces the Nonparametric Hamiltonian Monte Carlo (NP-HMC)
algorithm which generalises HMC to nonparametric models. Inputs to NP-HMC are a
new class of measurable functions called "tree representable", which serve as a
language-independent representation of the density functions of probabilistic
programs in a universal PPL. We provide a correctness proof of NP-HMC, and
empirically demonstrate significant performance improvements over existing
approaches on several nonparametric examples.
Related papers
- Lifted Model Construction without Normalisation: A Vectorised Approach to Exploit Symmetries in Factor Graphs [3.1045268505532566]
We find that the current state-of-the-art algorithm to construct a lifted representation in form of a parametric factor graph misses symmetries between factors that are exchangeable but scaled differently.
We propose a generalisation of the advanced colour passing (ACP) algorithm, which is the state of the art to construct a parametric factor graph.
Our proposed algorithm allows for potentials of factors to be scaled arbitrarily and efficiently detects more symmetries than the original ACP algorithm.
arXiv Detail & Related papers (2024-11-18T16:59:44Z) - Online Variational Sequential Monte Carlo [49.97673761305336]
We build upon the variational sequential Monte Carlo (VSMC) method, which provides computationally efficient and accurate model parameter estimation and Bayesian latent-state inference.
Online VSMC is capable of performing efficiently, entirely on-the-fly, both parameter estimation and particle proposal adaptation.
arXiv Detail & Related papers (2023-12-19T21:45:38Z) - Nonparametric Involutive Markov Chain Monte Carlo [6.445605125467574]
We show that NP-iMCMC can generalise numerous existing iMCMC algorithms to work on nonparametric models.
Applying our method to the recently proposed Nonparametric HMC, an instance of (Multiple Step) NP-iMCMC, we have constructed several nonparametric extensions.
arXiv Detail & Related papers (2022-11-02T13:21:52Z) - On MCMC for variationally sparse Gaussian processes: A pseudo-marginal
approach [0.76146285961466]
Gaussian processes (GPs) are frequently used in machine learning and statistics to construct powerful models.
We propose a pseudo-marginal (PM) scheme that offers exact inference as well as computational gains through doubly estimators for the likelihood and large datasets.
arXiv Detail & Related papers (2021-03-04T20:48:29Z) - Parallel Stochastic Mirror Descent for MDPs [72.75921150912556]
We consider the problem of learning the optimal policy for infinite-horizon Markov decision processes (MDPs)
Some variant of Mirror Descent is proposed for convex programming problems with Lipschitz-continuous functionals.
We analyze this algorithm in a general case and obtain an estimate of the convergence rate that does not accumulate errors during the operation of the method.
arXiv Detail & Related papers (2021-02-27T19:28:39Z) - Probabilistic Generating Circuits [50.98473654244851]
We propose probabilistic generating circuits (PGCs) for their efficient representation.
PGCs are not just a theoretical framework that unifies vastly different existing models, but also show huge potential in modeling realistic data.
We exhibit a simple class of PGCs that are not trivially subsumed by simple combinations of PCs and DPPs, and obtain competitive performance on a suite of density estimation benchmarks.
arXiv Detail & Related papers (2021-02-19T07:06:53Z) - Gaussian Process Latent Class Choice Models [7.992550355579791]
We present a non-parametric class of probabilistic machine learning within discrete choice models (DCMs)
The proposed model would assign individuals probabilistically to behaviorally homogeneous clusters (latent classes) using GPs.
The model is tested on two different mode choice applications and compared against different LCCM benchmarks.
arXiv Detail & Related papers (2021-01-28T19:56:42Z) - Community Detection in the Stochastic Block Model by Mixed Integer
Programming [3.8073142980733]
Degree-Corrected Block Model (DCSBM) is a popular model to generate random graphs with community structure given an expected degree sequence.
Standard approach of community detection based on the DCSBM is to search for the model parameters that are the most likely to have produced the observed network data through maximum likelihood estimation (MLE)
We present mathematical programming formulations and exact solution methods that can provably find the model parameters and community assignments of maximum likelihood given an observed graph.
arXiv Detail & Related papers (2021-01-26T22:04:40Z) - Efficient semidefinite-programming-based inference for binary and
multi-class MRFs [83.09715052229782]
We propose an efficient method for computing the partition function or MAP estimate in a pairwise MRF.
We extend semidefinite relaxations from the typical binary MRF to the full multi-class setting, and develop a compact semidefinite relaxation that can again be solved efficiently using the solver.
arXiv Detail & Related papers (2020-12-04T15:36:29Z) - Sinkhorn Natural Gradient for Generative Models [125.89871274202439]
We propose a novel Sinkhorn Natural Gradient (SiNG) algorithm which acts as a steepest descent method on the probability space endowed with the Sinkhorn divergence.
We show that the Sinkhorn information matrix (SIM), a key component of SiNG, has an explicit expression and can be evaluated accurately in complexity that scales logarithmically.
In our experiments, we quantitatively compare SiNG with state-of-the-art SGD-type solvers on generative tasks to demonstrate its efficiency and efficacy of our method.
arXiv Detail & Related papers (2020-11-09T02:51:17Z) - A Functional Perspective on Learning Symmetric Functions with Neural
Networks [48.80300074254758]
We study the learning and representation of neural networks defined on measures.
We establish approximation and generalization bounds under different choices of regularization.
The resulting models can be learned efficiently and enjoy generalization guarantees that extend across input sizes.
arXiv Detail & Related papers (2020-08-16T16:34:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.