Deep Ensemble as a Gaussian Process Approximate Posterior
- URL: http://arxiv.org/abs/2205.00163v1
- Date: Sat, 30 Apr 2022 05:25:44 GMT
- Title: Deep Ensemble as a Gaussian Process Approximate Posterior
- Authors: Zhijie Deng, Feng Zhou, Jianfei Chen, Guoqiang Wu, Jun Zhu
- Abstract summary: Deep Ensemble (DE) is an effective alternative to Bayesian neural networks for uncertainty quantification in deep learning.
We propose a refinement of DE where the functional inconsistency is explicitly characterized and tuned w.r.t. the training data and certain priori beliefs.
Our approach consumes only marginally added training cost than the standard DE, but achieves better uncertainty quantification than DE and its variants across diverse scenarios.
- Score: 32.617549688656354
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep Ensemble (DE) is an effective alternative to Bayesian neural networks
for uncertainty quantification in deep learning. The uncertainty of DE is
usually conveyed by the functional inconsistency among the ensemble members,
say, the disagreement among their predictions. Yet, the functional
inconsistency stems from unmanageable randomness and may easily collapse in
specific cases. To render the uncertainty of DE reliable, we propose a
refinement of DE where the functional inconsistency is explicitly
characterized, and further tuned w.r.t. the training data and certain priori
beliefs. Specifically, we describe the functional inconsistency with the
empirical covariance of the functions dictated by ensemble members, which,
along with the mean, define a Gaussian process (GP). Then, with specific priori
uncertainty imposed, we maximize functional evidence lower bound to make the GP
specified by DE approximate the Bayesian posterior. In this way, we relate DE
to Bayesian inference to enjoy reliable Bayesian uncertainty. Moreover, we
provide strategies to make the training efficient. Our approach consumes only
marginally added training cost than the standard DE, but achieves better
uncertainty quantification than DE and its variants across diverse scenarios.
Related papers
- Generalized Gaussian Temporal Difference Error for Uncertainty-aware Reinforcement Learning [0.19418036471925312]
We introduce a novel framework for generalized Gaussian error modeling in deep reinforcement learning.
Our framework enhances the flexibility of error distribution modeling by incorporating additional higher-order moment, particularly kurtosis.
arXiv Detail & Related papers (2024-08-05T08:12:25Z) - Bayesian meta learning for trustworthy uncertainty quantification [3.683202928838613]
We propose, Trust-Bayes, a novel optimization framework for Bayesian meta learning.
We characterize the lower bounds of the probabilities of the ground truth being captured by the specified intervals.
We analyze the sample complexity with respect to the feasible probability for trustworthy uncertainty quantification.
arXiv Detail & Related papers (2024-07-27T15:56:12Z) - Information-Theoretic Safe Bayesian Optimization [59.758009422067005]
We consider a sequential decision making task, where the goal is to optimize an unknown function without evaluating parameters that violate an unknown (safety) constraint.
Most current methods rely on a discretization of the domain and cannot be directly extended to the continuous case.
We propose an information-theoretic safe exploration criterion that directly exploits the GP posterior to identify the most informative safe parameters to evaluate.
arXiv Detail & Related papers (2024-02-23T14:31:10Z) - Decomposing Uncertainty for Large Language Models through Input Clarification Ensembling [69.83976050879318]
In large language models (LLMs), identifying sources of uncertainty is an important step toward improving reliability, trustworthiness, and interpretability.
In this paper, we introduce an uncertainty decomposition framework for LLMs, called input clarification ensembling.
Our approach generates a set of clarifications for the input, feeds them into an LLM, and ensembles the corresponding predictions.
arXiv Detail & Related papers (2023-11-15T05:58:35Z) - Likelihood Ratio Confidence Sets for Sequential Decision Making [51.66638486226482]
We revisit the likelihood-based inference principle and propose to use likelihood ratios to construct valid confidence sequences.
Our method is especially suitable for problems with well-specified likelihoods.
We show how to provably choose the best sequence of estimators and shed light on connections to online convex optimization.
arXiv Detail & Related papers (2023-11-08T00:10:21Z) - Model-Based Uncertainty in Value Functions [89.31922008981735]
We focus on characterizing the variance over values induced by a distribution over MDPs.
Previous work upper bounds the posterior variance over values by solving a so-called uncertainty Bellman equation.
We propose a new uncertainty Bellman equation whose solution converges to the true posterior variance over values.
arXiv Detail & Related papers (2023-02-24T09:18:27Z) - Data-Driven Influence Functions for Optimization-Based Causal Inference [105.5385525290466]
We study a constructive algorithm that approximates Gateaux derivatives for statistical functionals by finite differencing.
We study the case where probability distributions are not known a priori but need to be estimated from data.
arXiv Detail & Related papers (2022-08-29T16:16:22Z) - On Uncertainty, Tempering, and Data Augmentation in Bayesian
Classification [47.13680267076843]
We show that explicitly accounting for aleatoric uncertainty significantly improves the performance of Bayesian neural networks.
We find that a cold posterior, tempered by a power greater than one, often more honestly reflects our beliefs about aleatoric uncertainty than no tempering.
arXiv Detail & Related papers (2022-03-30T17:17:50Z) - Dense Uncertainty Estimation via an Ensemble-based Conditional Latent
Variable Model [68.34559610536614]
We argue that the aleatoric uncertainty is an inherent attribute of the data and can only be correctly estimated with an unbiased oracle model.
We propose a new sampling and selection strategy at train time to approximate the oracle model for aleatoric uncertainty estimation.
Our results show that our solution achieves both accurate deterministic results and reliable uncertainty estimation.
arXiv Detail & Related papers (2021-11-22T08:54:10Z) - Scalable Uncertainty for Computer Vision with Functional Variational
Inference [18.492485304537134]
We leverage the formulation of variational inference in function space.
We obtain predictive uncertainty estimates at the cost of a single forward pass through any chosen CNN architecture.
We propose numerically efficient algorithms which enable fast training in the context of high-dimensional tasks.
arXiv Detail & Related papers (2020-03-06T19:09:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.