Benchmarking common uncertainty estimation methods with
histopathological images under domain shift and label noise
- URL: http://arxiv.org/abs/2301.01054v2
- Date: Thu, 6 Jul 2023 10:38:54 GMT
- Title: Benchmarking common uncertainty estimation methods with
histopathological images under domain shift and label noise
- Authors: Hendrik A. Mehrtens, Alexander Kurz, Tabea-Clara Bucher, Titus J.
Brinker
- Abstract summary: In high-risk environments, deep learning models need to be able to judge their uncertainty and reject inputs when there is a significant chance of misclassification.
We conduct a rigorous evaluation of the most commonly used uncertainty and robustness methods for the classification of Whole Slide Images.
We observe that ensembles of methods generally lead to better uncertainty estimates as well as an increased robustness towards domain shifts and label noise.
- Score: 62.997667081978825
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: In the past years, deep learning has seen an increase in usage in the domain
of histopathological applications. However, while these approaches have shown
great potential, in high-risk environments deep learning models need to be able
to judge their uncertainty and be able to reject inputs when there is a
significant chance of misclassification. In this work, we conduct a rigorous
evaluation of the most commonly used uncertainty and robustness methods for the
classification of Whole Slide Images, with a focus on the task of selective
classification, where the model should reject the classification in situations
in which it is uncertain. We conduct our experiments on tile-level under the
aspects of domain shift and label noise, as well as on slide-level. In our
experiments, we compare Deep Ensembles, Monte-Carlo Dropout, Stochastic
Variational Inference, Test-Time Data Augmentation as well as ensembles of the
latter approaches. We observe that ensembles of methods generally lead to
better uncertainty estimates as well as an increased robustness towards domain
shifts and label noise, while contrary to results from classical computer
vision benchmarks no systematic gain of the other methods can be shown. Across
methods, a rejection of the most uncertain samples reliably leads to a
significant increase in classification accuracy on both in-distribution as well
as out-of-distribution data. Furthermore, we conduct experiments comparing
these methods under varying conditions of label noise. Lastly, we publish our
code framework to facilitate further research on uncertainty estimation on
histopathological data.
Related papers
- One step closer to unbiased aleatoric uncertainty estimation [71.55174353766289]
We propose a new estimation method by actively de-noising the observed data.
By conducting a broad range of experiments, we demonstrate that our proposed approach provides a much closer approximation to the actual data uncertainty than the standard method.
arXiv Detail & Related papers (2023-12-16T14:59:11Z) - Evidential Uncertainty Quantification: A Variance-Based Perspective [0.43536523356694407]
We adapt the variance-based approach from regression to classification, quantifying classification uncertainty at the class level.
Experiments on cross-domain datasets are conducted to illustrate that the variance-based approach not only results in similar accuracy as the entropy-based one in active domain adaptation.
arXiv Detail & Related papers (2023-11-19T16:33:42Z) - Semi-Supervised Deep Regression with Uncertainty Consistency and
Variational Model Ensembling via Bayesian Neural Networks [31.67508478764597]
We propose a novel approach to semi-supervised regression, namely Uncertainty-Consistent Variational Model Ensembling (UCVME)
Our consistency loss significantly improves uncertainty estimates and allows higher quality pseudo-labels to be assigned greater importance under heteroscedastic regression.
Experiments show that our method outperforms state-of-the-art alternatives on different tasks and can be competitive with supervised methods that use full labels.
arXiv Detail & Related papers (2023-02-15T10:40:51Z) - Systematic Evaluation of Predictive Fairness [60.0947291284978]
Mitigating bias in training on biased datasets is an important open problem.
We examine the performance of various debiasing methods across multiple tasks.
We find that data conditions have a strong influence on relative model performance.
arXiv Detail & Related papers (2022-10-17T05:40:13Z) - Scalable Marginal Likelihood Estimation for Model Selection in Deep
Learning [78.83598532168256]
Marginal-likelihood based model-selection is rarely used in deep learning due to estimation difficulties.
Our work shows that marginal likelihoods can improve generalization and be useful when validation data is unavailable.
arXiv Detail & Related papers (2021-04-11T09:50:24Z) - Learning Noise Transition Matrix from Only Noisy Labels via Total
Variation Regularization [88.91872713134342]
We propose a theoretically grounded method that can estimate the noise transition matrix and learn a classifier simultaneously.
We show the effectiveness of the proposed method through experiments on benchmark and real-world datasets.
arXiv Detail & Related papers (2021-02-04T05:09:18Z) - Minimax Active Learning [61.729667575374606]
Active learning aims to develop label-efficient algorithms by querying the most representative samples to be labeled by a human annotator.
Current active learning techniques either rely on model uncertainty to select the most uncertain samples or use clustering or reconstruction to choose the most diverse set of unlabeled examples.
We develop a semi-supervised minimax entropy-based active learning algorithm that leverages both uncertainty and diversity in an adversarial manner.
arXiv Detail & Related papers (2020-12-18T19:03:40Z) - Empirical Frequentist Coverage of Deep Learning Uncertainty
Quantification Procedures [13.890139530120164]
We provide the first large scale evaluation of the empirical frequentist coverage properties of uncertainty quantification techniques.
We find that, in general, some methods do achieve desirable coverage properties on in distribution samples, but that coverage is not maintained on out-of-distribution data.
arXiv Detail & Related papers (2020-10-06T21:22:46Z) - Identifying Causal-Effect Inference Failure with Uncertainty-Aware
Models [41.53326337725239]
We introduce a practical approach for integrating uncertainty estimation into a class of state-of-the-art neural network methods.
We show that our methods enable us to deal gracefully with situations of "no-overlap", common in high-dimensional data.
We show that correctly modeling uncertainty can keep us from giving overconfident and potentially harmful recommendations.
arXiv Detail & Related papers (2020-07-01T00:37:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.