A Heavy-Tailed Algebra for Probabilistic Programming
- URL: http://arxiv.org/abs/2306.09262v1
- Date: Thu, 15 Jun 2023 16:37:36 GMT
- Title: A Heavy-Tailed Algebra for Probabilistic Programming
- Authors: Feynman Liang, Liam Hodgkinson, Michael W. Mahoney
- Abstract summary: We propose a systematic approach for analyzing the tails of random variables.
We show how this approach can be used during the static analysis (before drawing samples) pass of a probabilistic programming language compiler.
Our empirical results confirm that inference algorithms that leverage our heavy-tailed algebra attain superior performance across a number of density modeling and variational inference tasks.
- Score: 53.32246823168763
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Despite the successes of probabilistic models based on passing noise through
neural networks, recent work has identified that such methods often fail to
capture tail behavior accurately, unless the tails of the base distribution are
appropriately calibrated. To overcome this deficiency, we propose a systematic
approach for analyzing the tails of random variables, and we illustrate how
this approach can be used during the static analysis (before drawing samples)
pass of a probabilistic programming language compiler. To characterize how the
tails change under various operations, we develop an algebra which acts on a
three-parameter family of tail asymptotics and which is based on the
generalized Gamma distribution. Our algebraic operations are closed under
addition and multiplication; they are capable of distinguishing sub-Gaussians
with differing scales; and they handle ratios sufficiently well to reproduce
the tails of most important statistical distributions directly from their
definitions. Our empirical results confirm that inference algorithms that
leverage our heavy-tailed algebra attain superior performance across a number
of density modeling and variational inference tasks.
Related papers
- Variational Elliptical Processes [1.5703073293718952]
We present elliptical processes, a family of non-parametric probabilistic models that subsume processes and Student's posterior processes.
We parameterize this mixture distribution as a spline normalizing flow, which we train using variational inference.
The proposed form of the variational posterior enables a sparse variational elliptical process applicable to large-scale problems.
arXiv Detail & Related papers (2023-11-21T12:26:14Z) - Robust Gaussian Process Regression with Huber Likelihood [2.7184224088243365]
We propose a robust process model in the Gaussian process framework with the likelihood of observed data expressed as the Huber probability distribution.
The proposed model employs weights based on projection statistics to scale residuals and bound the influence of vertical outliers and bad leverage points on the latent functions estimates.
arXiv Detail & Related papers (2023-01-19T02:59:33Z) - Distributional Gradient Boosting Machines [77.34726150561087]
Our framework is based on XGBoost and LightGBM.
We show that our framework achieves state-of-the-art forecast accuracy.
arXiv Detail & Related papers (2022-04-02T06:32:19Z) - Probabilistic Gradient Boosting Machines for Large-Scale Probabilistic
Regression [51.770998056563094]
Probabilistic Gradient Boosting Machines (PGBM) is a method to create probabilistic predictions with a single ensemble of decision trees.
We empirically demonstrate the advantages of PGBM compared to existing state-of-the-art methods.
arXiv Detail & Related papers (2021-06-03T08:32:13Z) - Scalable Cross Validation Losses for Gaussian Process Models [22.204619587725208]
We use Polya-Gamma auxiliary variables and variational inference to accommodate binary and multi-class classification.
We find that our method offers fast training and excellent predictive performance.
arXiv Detail & Related papers (2021-05-24T21:01:47Z) - Flexible mean field variational inference using mixtures of
non-overlapping exponential families [6.599344783327053]
I show that using standard mean field variational inference can fail to produce sensible results for models with sparsity-inducing priors.
I show that any mixture of a diffuse exponential family and a point mass at zero to model sparsity forms an exponential family.
arXiv Detail & Related papers (2020-10-14T01:46:56Z) - A Class of Conjugate Priors for Multinomial Probit Models which Includes
the Multivariate Normal One [0.3553493344868413]
We show that the entire class of unified skew-normal (SUN) distributions is conjugate to several multinomial probit models.
We improve upon state-of-the-art solutions for posterior inference and classification.
arXiv Detail & Related papers (2020-07-14T10:08:23Z) - Multiplicative noise and heavy tails in stochastic optimization [62.993432503309485]
empirical optimization is central to modern machine learning, but its role in its success is still unclear.
We show that it commonly arises in parameters of discrete multiplicative noise due to variance.
A detailed analysis is conducted in which we describe on key factors, including recent step size, and data, all exhibit similar results on state-of-the-art neural network models.
arXiv Detail & Related papers (2020-06-11T09:58:01Z) - Efficiently Sampling Functions from Gaussian Process Posteriors [76.94808614373609]
We propose an easy-to-use and general-purpose approach for fast posterior sampling.
We demonstrate how decoupled sample paths accurately represent Gaussian process posteriors at a fraction of the usual cost.
arXiv Detail & Related papers (2020-02-21T14:03:16Z) - Bayesian Deep Learning and a Probabilistic Perspective of Generalization [56.69671152009899]
We show that deep ensembles provide an effective mechanism for approximate Bayesian marginalization.
We also propose a related approach that further improves the predictive distribution by marginalizing within basins of attraction.
arXiv Detail & Related papers (2020-02-20T15:13:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.