A Deeper Look into Aleatoric and Epistemic Uncertainty Disentanglement
- URL: http://arxiv.org/abs/2204.09308v1
- Date: Wed, 20 Apr 2022 08:41:37 GMT
- Title: A Deeper Look into Aleatoric and Epistemic Uncertainty Disentanglement
- Authors: Matias Valdenegro-Toro and Daniel Saromo
- Abstract summary: In this paper, we generalize methods to produce disentangled uncertainties to work with different uncertainty quantification methods.
We show that there is an interaction between learning aleatoric and epistemic uncertainty, which is unexpected and violates assumptions on aleatoric uncertainty.
We expect that our formulation and results help practitioners and researchers choose uncertainty methods and expand the use of disentangled uncertainties.
- Score: 7.6146285961466
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Neural networks are ubiquitous in many tasks, but trusting their predictions
is an open issue. Uncertainty quantification is required for many applications,
and disentangled aleatoric and epistemic uncertainties are best. In this paper,
we generalize methods to produce disentangled uncertainties to work with
different uncertainty quantification methods, and evaluate their capability to
produce disentangled uncertainties. Our results show that: there is an
interaction between learning aleatoric and epistemic uncertainty, which is
unexpected and violates assumptions on aleatoric uncertainty, some methods like
Flipout produce zero epistemic uncertainty, aleatoric uncertainty is unreliable
in the out-of-distribution setting, and Ensembles provide overall the best
disentangling quality. We also explore the error produced by the number of
samples hyper-parameter in the sampling softmax function, recommending N > 100
samples. We expect that our formulation and results help practitioners and
researchers choose uncertainty methods and expand the use of disentangled
uncertainties, as well as motivate additional research into this topic.
Related papers
- How disentangled are your classification uncertainties? [6.144680854063938]
Uncertainty Quantification in Machine Learning has progressed to predicting the source of uncertainty in a prediction.
This work proposes a set of experiments to evaluate disentanglement of aleatoric and epistemic uncertainty.
arXiv Detail & Related papers (2024-08-22T07:42:43Z) - A unified uncertainty-aware exploration: Combining epistemic and
aleatory uncertainty [21.139502047972684]
We propose an algorithm that quantifies the combined effect of aleatory and epistemic uncertainty for risk-sensitive exploration.
Our method builds on a novel extension of distributional RL that estimates a parameterized return distribution.
Experimental results on tasks with exploration and risk challenges show that our method outperforms alternative approaches.
arXiv Detail & Related papers (2024-01-05T17:39:00Z) - One step closer to unbiased aleatoric uncertainty estimation [71.55174353766289]
We propose a new estimation method by actively de-noising the observed data.
By conducting a broad range of experiments, we demonstrate that our proposed approach provides a much closer approximation to the actual data uncertainty than the standard method.
arXiv Detail & Related papers (2023-12-16T14:59:11Z) - Decomposing Uncertainty for Large Language Models through Input Clarification Ensembling [69.83976050879318]
In large language models (LLMs), identifying sources of uncertainty is an important step toward improving reliability, trustworthiness, and interpretability.
In this paper, we introduce an uncertainty decomposition framework for LLMs, called input clarification ensembling.
Our approach generates a set of clarifications for the input, feeds them into an LLM, and ensembles the corresponding predictions.
arXiv Detail & Related papers (2023-11-15T05:58:35Z) - How certain are your uncertainties? [0.3655021726150368]
Measures of uncertainty in the output of a deep learning method are useful in several ways.
This work investigates the stability of these uncertainty measurements, in terms of both magnitude and spatial pattern.
arXiv Detail & Related papers (2022-03-01T05:25:02Z) - Bayesian autoencoders with uncertainty quantification: Towards
trustworthy anomaly detection [78.24964622317634]
In this work, the formulation of Bayesian autoencoders (BAEs) is adopted to quantify the total anomaly uncertainty.
To evaluate the quality of uncertainty, we consider the task of classifying anomalies with the additional option of rejecting predictions of high uncertainty.
Our experiments demonstrate the effectiveness of the BAE and total anomaly uncertainty on a set of benchmark datasets and two real datasets for manufacturing.
arXiv Detail & Related papers (2022-02-25T12:20:04Z) - Dense Uncertainty Estimation via an Ensemble-based Conditional Latent
Variable Model [68.34559610536614]
We argue that the aleatoric uncertainty is an inherent attribute of the data and can only be correctly estimated with an unbiased oracle model.
We propose a new sampling and selection strategy at train time to approximate the oracle model for aleatoric uncertainty estimation.
Our results show that our solution achieves both accurate deterministic results and reliable uncertainty estimation.
arXiv Detail & Related papers (2021-11-22T08:54:10Z) - Deterministic Neural Networks with Appropriate Inductive Biases Capture
Epistemic and Aleatoric Uncertainty [91.01037972035635]
We show that a single softmax neural net with minimal changes can beat the uncertainty predictions of Deep Ensembles.
We study why, and show that with the right inductive biases, softmax neural nets trained with maximum likelihood reliably capture uncertainty through the feature-space density.
arXiv Detail & Related papers (2021-02-23T09:44:09Z) - The Hidden Uncertainty in a Neural Networks Activations [105.4223982696279]
The distribution of a neural network's latent representations has been successfully used to detect out-of-distribution (OOD) data.
This work investigates whether this distribution correlates with a model's epistemic uncertainty, thus indicating its ability to generalise to novel inputs.
arXiv Detail & Related papers (2020-12-05T17:30:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.