Deep interpretable ensembles
- URL: http://arxiv.org/abs/2205.12729v1
- Date: Wed, 25 May 2022 12:39:39 GMT
- Title: Deep interpretable ensembles
- Authors: Lucas Kook, Andrea G\"otschi, Philipp FM Baumann, Torsten Hothorn,
Beate Sick
- Abstract summary: In deep ensembling, the individual models are usually black box neural networks, or recently, partially interpretable semi-structured deep transformation models.
We propose a novel transformation ensemble which aggregates probabilistic predictions with the guarantee to preserve interpretability and yield uniformly better predictions than the ensemble members on average.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Ensembles improve prediction performance and allow uncertainty quantification
by aggregating predictions from multiple models. In deep ensembling, the
individual models are usually black box neural networks, or recently, partially
interpretable semi-structured deep transformation models. However,
interpretability of the ensemble members is generally lost upon aggregation.
This is a crucial drawback of deep ensembles in high-stake decision fields, in
which interpretable models are desired. We propose a novel transformation
ensemble which aggregates probabilistic predictions with the guarantee to
preserve interpretability and yield uniformly better predictions than the
ensemble members on average. Transformation ensembles are tailored towards
interpretable deep transformation models but are applicable to a wider range of
probabilistic neural networks. In experiments on several publicly available
data sets, we demonstrate that transformation ensembles perform on par with
classical deep ensembles in terms of prediction performance, discrimination,
and calibration. In addition, we demonstrate how transformation ensembles
quantify both aleatoric and epistemic uncertainty, and produce minimax optimal
predictions under certain conditions.
Related papers
- Mixup Regularization: A Probabilistic Perspective [11.501663622998697]
We introduce a novel framework for mixup regularization based on probabilistic fusion.
For data distributed according to a member of the exponential family, we show that likelihood functions can be analytically fused.
We propose an extension of probabilistic mixup, which allows for fusion of inputs at an arbitrary intermediate layer of the neural network.
arXiv Detail & Related papers (2025-02-19T15:39:14Z) - Can Bayesian Neural Networks Make Confident Predictions? [0.0]
We show that under a discretized prior for the inner layer weights, we can exactly characterize the posterior predictive distribution as a Gaussian mixture.
We also characterize the capacity of a model to "learn from data" by evaluating contraction of the posterior predictive in different scaling regimes.
arXiv Detail & Related papers (2025-01-20T22:36:28Z) - Dynamic Post-Hoc Neural Ensemblers [55.15643209328513]
In this study, we explore employing neural networks as ensemble methods.
Motivated by the risk of learning low-diversity ensembles, we propose regularizing the model by randomly dropping base model predictions.
We demonstrate this approach lower bounds the diversity within the ensemble, reducing overfitting and improving generalization capabilities.
arXiv Detail & Related papers (2024-10-06T15:25:39Z) - Revisiting Optimism and Model Complexity in the Wake of Overparameterized Machine Learning [6.278498348219108]
We revisit model complexity from first principles, by first reinterpreting and then extending the classical statistical concept of (effective) degrees of freedom.
We demonstrate the utility of our proposed complexity measures through a mix of conceptual arguments, theory, and experiments.
arXiv Detail & Related papers (2024-10-02T06:09:57Z) - Towards Generalizable and Interpretable Motion Prediction: A Deep
Variational Bayes Approach [54.429396802848224]
This paper proposes an interpretable generative model for motion prediction with robust generalizability to out-of-distribution cases.
For interpretability, the model achieves the target-driven motion prediction by estimating the spatial distribution of long-term destinations.
Experiments on motion prediction datasets validate that the fitted model can be interpretable and generalizable.
arXiv Detail & Related papers (2024-03-10T04:16:04Z) - On the Out-of-Distribution Coverage of Combining Split Conformal
Prediction and Bayesian Deep Learning [1.131316248570352]
We focus on combining Bayesian deep learning with split conformal prediction and how this combination effects out-of-distribution coverage.
Our results suggest that combining Bayesian deep learning models with split conformal prediction can, in some cases, cause unintended consequences such as reducing out-of-distribution coverage.
arXiv Detail & Related papers (2023-11-21T15:50:37Z) - Structured Radial Basis Function Network: Modelling Diversity for
Multiple Hypotheses Prediction [51.82628081279621]
Multi-modal regression is important in forecasting nonstationary processes or with a complex mixture of distributions.
A Structured Radial Basis Function Network is presented as an ensemble of multiple hypotheses predictors for regression problems.
It is proved that this structured model can efficiently interpolate this tessellation and approximate the multiple hypotheses target distribution.
arXiv Detail & Related papers (2023-09-02T01:27:53Z) - Aggregating distribution forecasts from deep ensembles [0.0]
We propose a general quantile aggregation framework for deep ensembles.
We show that combining forecast distributions from deep ensembles can substantially improve the predictive performance.
arXiv Detail & Related papers (2022-04-05T15:42:51Z) - Distributional Gradient Boosting Machines [77.34726150561087]
Our framework is based on XGBoost and LightGBM.
We show that our framework achieves state-of-the-art forecast accuracy.
arXiv Detail & Related papers (2022-04-02T06:32:19Z) - Probabilistic electric load forecasting through Bayesian Mixture Density
Networks [70.50488907591463]
Probabilistic load forecasting (PLF) is a key component in the extended tool-chain required for efficient management of smart energy grids.
We propose a novel PLF approach, framed on Bayesian Mixture Density Networks.
To achieve reliable and computationally scalable estimators of the posterior distributions, both Mean Field variational inference and deep ensembles are integrated.
arXiv Detail & Related papers (2020-12-23T16:21:34Z) - Bayesian Deep Learning and a Probabilistic Perspective of Generalization [56.69671152009899]
We show that deep ensembles provide an effective mechanism for approximate Bayesian marginalization.
We also propose a related approach that further improves the predictive distribution by marginalizing within basins of attraction.
arXiv Detail & Related papers (2020-02-20T15:13:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.