Slice Sampling for General Completely Random Measures
- URL: http://arxiv.org/abs/2006.13925v2
- Date: Thu, 25 Jun 2020 07:30:01 GMT
- Title: Slice Sampling for General Completely Random Measures
- Authors: Peiyuan Zhu, Alexandre Bouchard-C\^ot\'e, and Trevor Campbell
- Abstract summary: We present a novel Markov chain Monte Carlo algorithm for posterior inference that adaptively sets the truncation level using auxiliary slice variables.
The efficacy of the proposed algorithm is evaluated on several popular nonparametric models.
- Score: 74.24975039689893
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Completely random measures provide a principled approach to creating flexible
unsupervised models, where the number of latent features is infinite and the
number of features that influence the data grows with the size of the data set.
Due to the infinity the latent features, posterior inference requires either
marginalization---resulting in dependence structures that prevent efficient
computation via parallelization and conjugacy---or finite truncation, which
arbitrarily limits the flexibility of the model. In this paper we present a
novel Markov chain Monte Carlo algorithm for posterior inference that
adaptively sets the truncation level using auxiliary slice variables, enabling
efficient, parallelized computation without sacrificing flexibility. In
contrast to past work that achieved this on a model-by-model basis, we provide
a general recipe that is applicable to the broad class of completely random
measure-based priors. The efficacy of the proposed algorithm is evaluated on
several popular nonparametric models, demonstrating a higher effective sample
size per second compared to algorithms using marginalization as well as a
higher predictive performance compared to models employing fixed truncations.
Related papers
- Scaling and renormalization in high-dimensional regression [72.59731158970894]
This paper presents a succinct derivation of the training and generalization performance of a variety of high-dimensional ridge regression models.
We provide an introduction and review of recent results on these topics, aimed at readers with backgrounds in physics and deep learning.
arXiv Detail & Related papers (2024-05-01T15:59:00Z) - Online Variational Sequential Monte Carlo [49.97673761305336]
We build upon the variational sequential Monte Carlo (VSMC) method, which provides computationally efficient and accurate model parameter estimation and Bayesian latent-state inference.
Online VSMC is capable of performing efficiently, entirely on-the-fly, both parameter estimation and particle proposal adaptation.
arXiv Detail & Related papers (2023-12-19T21:45:38Z) - Amortizing intractable inference in large language models [56.92471123778389]
We use amortized Bayesian inference to sample from intractable posterior distributions.
We empirically demonstrate that this distribution-matching paradigm of LLM fine-tuning can serve as an effective alternative to maximum-likelihood training.
As an important application, we interpret chain-of-thought reasoning as a latent variable modeling problem.
arXiv Detail & Related papers (2023-10-06T16:36:08Z) - Approximate Message Passing for the Matrix Tensor Product Model [8.206394018475708]
We propose and analyze an approximate message passing (AMP) algorithm for the matrix tensor product model.
Building upon an convergence theorem for non-separable functions, we prove a state evolution for non-separable functions.
We leverage this state evolution result to provide necessary and sufficient conditions for recovery of the signal of interest.
arXiv Detail & Related papers (2023-06-27T16:03:56Z) - A model-free feature selection technique of feature screening and random
forest based recursive feature elimination [0.0]
We propose a model-free feature selection method for ultra-high dimensional data with mass features.
We show that the proposed method is selection consistent and $L$ consistent under weak regularity conditions.
arXiv Detail & Related papers (2023-02-15T03:39:16Z) - Spike-and-Slab Generalized Additive Models and Scalable Algorithms for
High-Dimensional Data [0.0]
We propose hierarchical generalized additive models (GAMs) to accommodate high-dimensional data.
We consider the smoothing penalty for proper shrinkage of curve and separation of smoothing function linear and nonlinear spaces.
Two and deterministic algorithms, EM-Coordinate Descent and EM-Iterative Weighted Least Squares, are developed for different utilities.
arXiv Detail & Related papers (2021-10-27T14:11:13Z) - Group selection and shrinkage: Structured sparsity for semiparametric
additive models [0.0]
Sparse regression and classification estimators that respect group structures have application to an assortment of statistical and machine learning problems.
We develop a framework for fitting the nonparametric surface and present finite error models.
We demonstrate their efficacy in modeling foot and economic predictors using many predictors.
arXiv Detail & Related papers (2021-05-25T17:00:25Z) - Efficient Ensemble Model Generation for Uncertainty Estimation with
Bayesian Approximation in Segmentation [74.06904875527556]
We propose a generic and efficient segmentation framework to construct ensemble segmentation models.
In the proposed method, ensemble models can be efficiently generated by using the layer selection method.
We also devise a new pixel-wise uncertainty loss, which improves the predictive performance.
arXiv Detail & Related papers (2020-05-21T16:08:38Z) - Asymptotic Analysis of an Ensemble of Randomly Projected Linear
Discriminants [94.46276668068327]
In [1], an ensemble of randomly projected linear discriminants is used to classify datasets.
We develop a consistent estimator of the misclassification probability as an alternative to the computationally-costly cross-validation estimator.
We also demonstrate the use of our estimator for tuning the projection dimension on both real and synthetic data.
arXiv Detail & Related papers (2020-04-17T12:47:04Z) - Scaling Bayesian inference of mixed multinomial logit models to very
large datasets [9.442139459221785]
We propose an Amortized Variational Inference approach that leverages backpropagation, automatic differentiation and GPU-accelerated computation.
We show how normalizing flows can be used to increase the flexibility of the variational posterior approximations.
arXiv Detail & Related papers (2020-04-11T15:30:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.