Multivariate Density Estimation with Deep Neural Mixture Models
- URL: http://arxiv.org/abs/2012.03391v1
- Date: Sun, 6 Dec 2020 23:03:48 GMT
- Title: Multivariate Density Estimation with Deep Neural Mixture Models
- Authors: Edmondo Trentin (DIISM, University of Siena, Italy)
- Abstract summary: Deep neural networks (DNNs) have seldom been applied to density estimation.
This paper extends our previous work on Neural Mixture Densities (NMMs)
A maximum-likelihood (ML) algorithm for estimating Deep NMMs (DNMMs) is handed out.
The class of probability density functions that can be modeled to any degree of precision via DNMMs is formally defined.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Albeit worryingly underrated in the recent literature on machine learning in
general (and, on deep learning in particular), multivariate density estimation
is a fundamental task in many applications, at least implicitly, and still an
open issue. With a few exceptions, deep neural networks (DNNs) have seldom been
applied to density estimation, mostly due to the unsupervised nature of the
estimation task, and (especially) due to the need for constrained training
algorithms that ended up realizing proper probabilistic models that satisfy
Kolmogorov's axioms. Moreover, in spite of the well-known improvement in terms
of modeling capabilities yielded by mixture models over plain single-density
statistical estimators, no proper mixtures of multivariate DNN-based component
densities have been investigated so far. The paper fills this gap by extending
our previous work on Neural Mixture Densities (NMMs) to multivariate DNN
mixtures. A maximum-likelihood (ML) algorithm for estimating Deep NMMs (DNMMs)
is handed out, which satisfies numerically a combination of hard and soft
constraints aimed at ensuring satisfaction of Kolmogorov's axioms. The class of
probability density functions that can be modeled to any degree of precision
via DNMMs is formally defined. A procedure for the automatic selection of the
DNMM architecture, as well as of the hyperparameters for its ML training
algorithm, is presented (exploiting the probabilistic nature of the DNMM).
Experimental results on univariate and multivariate data are reported on,
corroborating the effectiveness of the approach and its superiority to the most
popular statistical estimation techniques.
Related papers
- MD-NOMAD: Mixture density nonlinear manifold decoder for emulating stochastic differential equations and uncertainty propagation [0.9208007322096533]
We propose a neural operator framework, termed mixture density nonlinear manifold decoder (MD-NOMAD) for simulators.
Our approach leverages an amalgamation of the pointwise operator learning neural architecture nonlinear manifold decoder (NOMAD) with mixture density-based methods to estimate conditional probability for output functions.
arXiv Detail & Related papers (2024-04-24T08:39:14Z) - A Multi-Head Ensemble Multi-Task Learning Approach for Dynamical
Computation Offloading [62.34538208323411]
We propose a multi-head ensemble multi-task learning (MEMTL) approach with a shared backbone and multiple prediction heads (PHs)
MEMTL outperforms benchmark methods in both the inference accuracy and mean square error without requiring additional training data.
arXiv Detail & Related papers (2023-09-02T11:01:16Z) - Deep Gaussian Mixture Ensembles [9.673093148930874]
This work introduces a novel probabilistic deep learning technique called deep Gaussian mixture ensembles (DGMEs)
DGMEs are capable of approximating complex probability distributions, such as heavy-tailed or multimodal distributions.
Our experimental results demonstrate that DGMEs outperform state-of-the-art uncertainty quantifying deep learning models in handling complex predictive densities.
arXiv Detail & Related papers (2023-06-12T16:53:38Z) - Single-model uncertainty quantification in neural network potentials
does not consistently outperform model ensembles [0.7499722271664145]
Neural networks (NNs) often assign high confidence to their predictions, even for points far out-of-distribution.
Uncertainty quantification (UQ) is a challenge when they are employed to model interatomic potentials in materials systems.
Differentiable UQ techniques can find new informative data and drive active learning loops for robust potentials.
arXiv Detail & Related papers (2023-05-02T19:41:17Z) - Validation Diagnostics for SBI algorithms based on Normalizing Flows [55.41644538483948]
This work proposes easy to interpret validation diagnostics for multi-dimensional conditional (posterior) density estimators based on NF.
It also offers theoretical guarantees based on results of local consistency.
This work should help the design of better specified models or drive the development of novel SBI-algorithms.
arXiv Detail & Related papers (2022-11-17T15:48:06Z) - Multi-fidelity Hierarchical Neural Processes [79.0284780825048]
Multi-fidelity surrogate modeling reduces the computational cost by fusing different simulation outputs.
We propose Multi-fidelity Hierarchical Neural Processes (MF-HNP), a unified neural latent variable model for multi-fidelity surrogate modeling.
We evaluate MF-HNP on epidemiology and climate modeling tasks, achieving competitive performance in terms of accuracy and uncertainty estimation.
arXiv Detail & Related papers (2022-06-10T04:54:13Z) - Information Theoretic Structured Generative Modeling [13.117829542251188]
A novel generative model framework called the structured generative model (SGM) is proposed that makes straightforward optimization possible.
The implementation employs a single neural network driven by an orthonormal input to a single white noise source adapted to learn an infinite Gaussian mixture model.
Preliminary results show that SGM significantly improves MINE estimation in terms of data efficiency and variance, conventional and variational Gaussian mixture models, as well as for training adversarial networks.
arXiv Detail & Related papers (2021-10-12T07:44:18Z) - Marginalizable Density Models [14.50261153230204]
We present a novel deep network architecture which provides closed form expressions for the probabilities, marginals and conditionals of any subset of the variables.
The model also allows for parallelized sampling with only a logarithmic dependence of the time complexity on the number of variables.
arXiv Detail & Related papers (2021-06-08T23:54:48Z) - Post-mortem on a deep learning contest: a Simpson's paradox and the
complementary roles of scale metrics versus shape metrics [61.49826776409194]
We analyze a corpus of models made publicly-available for a contest to predict the generalization accuracy of neural network (NN) models.
We identify what amounts to a Simpson's paradox: where "scale" metrics perform well overall but perform poorly on sub partitions of the data.
We present two novel shape metrics, one data-independent, and the other data-dependent, which can predict trends in the test accuracy of a series of NNs.
arXiv Detail & Related papers (2021-06-01T19:19:49Z) - Amortized Conditional Normalized Maximum Likelihood: Reliable Out of
Distribution Uncertainty Estimation [99.92568326314667]
We propose the amortized conditional normalized maximum likelihood (ACNML) method as a scalable general-purpose approach for uncertainty estimation.
Our algorithm builds on the conditional normalized maximum likelihood (CNML) coding scheme, which has minimax optimal properties according to the minimum description length principle.
We demonstrate that ACNML compares favorably to a number of prior techniques for uncertainty estimation in terms of calibration on out-of-distribution inputs.
arXiv Detail & Related papers (2020-11-05T08:04:34Z) - Large-scale Neural Solvers for Partial Differential Equations [48.7576911714538]
Solving partial differential equations (PDE) is an indispensable part of many branches of science as many processes can be modelled in terms of PDEs.
Recent numerical solvers require manual discretization of the underlying equation as well as sophisticated, tailored code for distributed computing.
We examine the applicability of continuous, mesh-free neural solvers for partial differential equations, physics-informed neural networks (PINNs)
We discuss the accuracy of GatedPINN with respect to analytical solutions -- as well as state-of-the-art numerical solvers, such as spectral solvers.
arXiv Detail & Related papers (2020-09-08T13:26:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.