Deep Gaussian Mixture Ensembles
- URL: http://arxiv.org/abs/2306.07235v1
- Date: Mon, 12 Jun 2023 16:53:38 GMT
- Title: Deep Gaussian Mixture Ensembles
- Authors: Yousef El-Laham, Niccol\`o Dalmasso, Elizabeth Fons, Svitlana
Vyetrenko
- Abstract summary: This work introduces a novel probabilistic deep learning technique called deep Gaussian mixture ensembles (DGMEs)
DGMEs are capable of approximating complex probability distributions, such as heavy-tailed or multimodal distributions.
Our experimental results demonstrate that DGMEs outperform state-of-the-art uncertainty quantifying deep learning models in handling complex predictive densities.
- Score: 9.673093148930874
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This work introduces a novel probabilistic deep learning technique called
deep Gaussian mixture ensembles (DGMEs), which enables accurate quantification
of both epistemic and aleatoric uncertainty. By assuming the data generating
process follows that of a Gaussian mixture, DGMEs are capable of approximating
complex probability distributions, such as heavy-tailed or multimodal
distributions. Our contributions include the derivation of an
expectation-maximization (EM) algorithm used for learning the model parameters,
which results in an upper-bound on the log-likelihood of training data over
that of standard deep ensembles. Additionally, the proposed EM training
procedure allows for learning of mixture weights, which is not commonly done in
ensembles. Our experimental results demonstrate that DGMEs outperform
state-of-the-art uncertainty quantifying deep learning models in handling
complex predictive densities.
Related papers
- MD-NOMAD: Mixture density nonlinear manifold decoder for emulating stochastic differential equations and uncertainty propagation [0.9208007322096533]
We propose a neural operator framework, termed mixture density nonlinear manifold decoder (MD-NOMAD) for simulators.
Our approach leverages an amalgamation of the pointwise operator learning neural architecture nonlinear manifold decoder (NOMAD) with mixture density-based methods to estimate conditional probability for output functions.
arXiv Detail & Related papers (2024-04-24T08:39:14Z) - Fusion of Gaussian Processes Predictions with Monte Carlo Sampling [61.31380086717422]
In science and engineering, we often work with models designed for accurate prediction of variables of interest.
Recognizing that these models are approximations of reality, it becomes desirable to apply multiple models to the same data and integrate their outcomes.
arXiv Detail & Related papers (2024-03-03T04:21:21Z) - Heterogeneous Multi-Task Gaussian Cox Processes [61.67344039414193]
We present a novel extension of multi-task Gaussian Cox processes for modeling heterogeneous correlated tasks jointly.
A MOGP prior over the parameters of the dedicated likelihoods for classification, regression and point process tasks can facilitate sharing of information between heterogeneous tasks.
We derive a mean-field approximation to realize closed-form iterative updates for estimating model parameters.
arXiv Detail & Related papers (2023-08-29T15:01:01Z) - Mixtures of Gaussian Process Experts with SMC$^2$ [0.4588028371034407]
mixtures of Gaussian process experts have been considered where data points are assigned to independent experts.
We construct a novel inference approach based on nested sequential Monte Carlo samplers to infer both the gating network and Gaussian process expert parameters.
arXiv Detail & Related papers (2022-08-26T18:20:14Z) - A Robust and Flexible EM Algorithm for Mixtures of Elliptical
Distributions with Missing Data [71.9573352891936]
This paper tackles the problem of missing data imputation for noisy and non-Gaussian data.
A new EM algorithm is investigated for mixtures of elliptical distributions with the property of handling potential missing data.
Experimental results on synthetic data demonstrate that the proposed algorithm is robust to outliers and can be used with non-Gaussian data.
arXiv Detail & Related papers (2022-01-28T10:01:37Z) - Information Theoretic Structured Generative Modeling [13.117829542251188]
A novel generative model framework called the structured generative model (SGM) is proposed that makes straightforward optimization possible.
The implementation employs a single neural network driven by an orthonormal input to a single white noise source adapted to learn an infinite Gaussian mixture model.
Preliminary results show that SGM significantly improves MINE estimation in terms of data efficiency and variance, conventional and variational Gaussian mixture models, as well as for training adversarial networks.
arXiv Detail & Related papers (2021-10-12T07:44:18Z) - Learning Gaussian Mixtures with Generalised Linear Models: Precise
Asymptotics in High-dimensions [79.35722941720734]
Generalised linear models for multi-class classification problems are one of the fundamental building blocks of modern machine learning tasks.
We prove exacts characterising the estimator in high-dimensions via empirical risk minimisation.
We discuss how our theory can be applied beyond the scope of synthetic data.
arXiv Detail & Related papers (2021-06-07T16:53:56Z) - Probabilistic electric load forecasting through Bayesian Mixture Density
Networks [70.50488907591463]
Probabilistic load forecasting (PLF) is a key component in the extended tool-chain required for efficient management of smart energy grids.
We propose a novel PLF approach, framed on Bayesian Mixture Density Networks.
To achieve reliable and computationally scalable estimators of the posterior distributions, both Mean Field variational inference and deep ensembles are integrated.
arXiv Detail & Related papers (2020-12-23T16:21:34Z) - Multivariate Density Estimation with Deep Neural Mixture Models [0.0]
Deep neural networks (DNNs) have seldom been applied to density estimation.
This paper extends our previous work on Neural Mixture Densities (NMMs)
A maximum-likelihood (ML) algorithm for estimating Deep NMMs (DNMMs) is handed out.
The class of probability density functions that can be modeled to any degree of precision via DNMMs is formally defined.
arXiv Detail & Related papers (2020-12-06T23:03:48Z) - A similarity-based Bayesian mixture-of-experts model [0.5156484100374058]
We present a new non-parametric mixture-of-experts model for multivariate regression problems.
Using a conditionally specified model, predictions for out-of-sample inputs are based on similarities to each observed data point.
Posterior inference is performed on the parameters of the mixture as well as the distance metric.
arXiv Detail & Related papers (2020-12-03T18:08:30Z) - Plug-And-Play Learned Gaussian-mixture Approximate Message Passing [71.74028918819046]
We propose a plug-and-play compressed sensing (CS) recovery algorithm suitable for any i.i.d. source prior.
Our algorithm builds upon Borgerding's learned AMP (LAMP), yet significantly improves it by adopting a universal denoising function within the algorithm.
Numerical evaluation shows that the L-GM-AMP algorithm achieves state-of-the-art performance without any knowledge of the source prior.
arXiv Detail & Related papers (2020-11-18T16:40:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.