Deep interpretable ensembles
- URL: http://arxiv.org/abs/2205.12729v1
- Date: Wed, 25 May 2022 12:39:39 GMT
- Title: Deep interpretable ensembles
- Authors: Lucas Kook, Andrea G\"otschi, Philipp FM Baumann, Torsten Hothorn,
Beate Sick
- Abstract summary: In deep ensembling, the individual models are usually black box neural networks, or recently, partially interpretable semi-structured deep transformation models.
We propose a novel transformation ensemble which aggregates probabilistic predictions with the guarantee to preserve interpretability and yield uniformly better predictions than the ensemble members on average.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Ensembles improve prediction performance and allow uncertainty quantification
by aggregating predictions from multiple models. In deep ensembling, the
individual models are usually black box neural networks, or recently, partially
interpretable semi-structured deep transformation models. However,
interpretability of the ensemble members is generally lost upon aggregation.
This is a crucial drawback of deep ensembles in high-stake decision fields, in
which interpretable models are desired. We propose a novel transformation
ensemble which aggregates probabilistic predictions with the guarantee to
preserve interpretability and yield uniformly better predictions than the
ensemble members on average. Transformation ensembles are tailored towards
interpretable deep transformation models but are applicable to a wider range of
probabilistic neural networks. In experiments on several publicly available
data sets, we demonstrate that transformation ensembles perform on par with
classical deep ensembles in terms of prediction performance, discrimination,
and calibration. In addition, we demonstrate how transformation ensembles
quantify both aleatoric and epistemic uncertainty, and produce minimax optimal
predictions under certain conditions.
Related papers
- BI-EqNO: Generalized Approximate Bayesian Inference with an Equivariant Neural Operator Framework [9.408644291433752]
We introduce BI-EqNO, an equivariant neural operator framework for generalized approximate Bayesian inference.
BI-EqNO transforms priors into posteriors on conditioned observation data through data-driven training.
We demonstrate BI-EqNO's utility through two examples: (1) as a generalized Gaussian process (gGP) for regression, and (2) as an ensemble neural filter (EnNF) for sequential data assimilation.
arXiv Detail & Related papers (2024-10-21T18:39:16Z) - Dynamic Post-Hoc Neural Ensemblers [55.15643209328513]
In this study, we explore employing neural networks as ensemble methods.
Motivated by the risk of learning low-diversity ensembles, we propose regularizing the model by randomly dropping base model predictions.
We demonstrate this approach lower bounds the diversity within the ensemble, reducing overfitting and improving generalization capabilities.
arXiv Detail & Related papers (2024-10-06T15:25:39Z) - Revisiting Optimism and Model Complexity in the Wake of Overparameterized Machine Learning [6.278498348219108]
We revisit model complexity from first principles, by first reinterpreting and then extending the classical statistical concept of (effective) degrees of freedom.
We demonstrate the utility of our proposed complexity measures through a mix of conceptual arguments, theory, and experiments.
arXiv Detail & Related papers (2024-10-02T06:09:57Z) - Towards Generalizable and Interpretable Motion Prediction: A Deep
Variational Bayes Approach [54.429396802848224]
This paper proposes an interpretable generative model for motion prediction with robust generalizability to out-of-distribution cases.
For interpretability, the model achieves the target-driven motion prediction by estimating the spatial distribution of long-term destinations.
Experiments on motion prediction datasets validate that the fitted model can be interpretable and generalizable.
arXiv Detail & Related papers (2024-03-10T04:16:04Z) - On the Out-of-Distribution Coverage of Combining Split Conformal
Prediction and Bayesian Deep Learning [1.131316248570352]
We focus on combining Bayesian deep learning with split conformal prediction and how this combination effects out-of-distribution coverage.
Our results suggest that combining Bayesian deep learning models with split conformal prediction can, in some cases, cause unintended consequences such as reducing out-of-distribution coverage.
arXiv Detail & Related papers (2023-11-21T15:50:37Z) - Structured Radial Basis Function Network: Modelling Diversity for
Multiple Hypotheses Prediction [51.82628081279621]
Multi-modal regression is important in forecasting nonstationary processes or with a complex mixture of distributions.
A Structured Radial Basis Function Network is presented as an ensemble of multiple hypotheses predictors for regression problems.
It is proved that this structured model can efficiently interpolate this tessellation and approximate the multiple hypotheses target distribution.
arXiv Detail & Related papers (2023-09-02T01:27:53Z) - Aggregating distribution forecasts from deep ensembles [0.0]
We study the question of how to aggregate distribution forecasts based on neural network-based approaches.
We show that combining forecast distributions can substantially improve the predictive performance.
We propose a general quantile aggregation framework for deep ensembles that shows superior performance compared to a linear combination of the forecast densities.
arXiv Detail & Related papers (2022-04-05T15:42:51Z) - Distributional Gradient Boosting Machines [77.34726150561087]
Our framework is based on XGBoost and LightGBM.
We show that our framework achieves state-of-the-art forecast accuracy.
arXiv Detail & Related papers (2022-04-02T06:32:19Z) - Probabilistic electric load forecasting through Bayesian Mixture Density
Networks [70.50488907591463]
Probabilistic load forecasting (PLF) is a key component in the extended tool-chain required for efficient management of smart energy grids.
We propose a novel PLF approach, framed on Bayesian Mixture Density Networks.
To achieve reliable and computationally scalable estimators of the posterior distributions, both Mean Field variational inference and deep ensembles are integrated.
arXiv Detail & Related papers (2020-12-23T16:21:34Z) - Set Prediction without Imposing Structure as Conditional Density
Estimation [40.86881969839325]
We propose an alternative to training via set losses by viewing learning as conditional density estimation.
Our framework fits deep energy-based models and approximates the intractable likelihood with gradient-guided sampling.
Our approach is competitive with previous set prediction models on standard benchmarks.
arXiv Detail & Related papers (2020-10-08T16:49:16Z) - Bayesian Deep Learning and a Probabilistic Perspective of Generalization [56.69671152009899]
We show that deep ensembles provide an effective mechanism for approximate Bayesian marginalization.
We also propose a related approach that further improves the predictive distribution by marginalizing within basins of attraction.
arXiv Detail & Related papers (2020-02-20T15:13:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.