On the calibration of neural networks for histological slide-level
classification
- URL: http://arxiv.org/abs/2312.09719v1
- Date: Fri, 15 Dec 2023 11:46:29 GMT
- Title: On the calibration of neural networks for histological slide-level
classification
- Authors: Alexander Kurz, Hendrik A. Mehrtens, Tabea-Clara Bucher, Titus J.
Brinker
- Abstract summary: We compare three neural network architectures that combine feature representations on patch-level to a slide-level prediction with respect to their classification performance.
We observe that Transformers lead to good results in terms of classification performance and calibration.
- Score: 47.99822253865054
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Deep Neural Networks have shown promising classification performance when
predicting certain biomarkers from Whole Slide Images in digital pathology.
However, the calibration of the networks' output probabilities is often not
evaluated. Communicating uncertainty by providing reliable confidence scores is
of high relevance in the medical context. In this work, we compare three neural
network architectures that combine feature representations on patch-level to a
slide-level prediction with respect to their classification performance and
evaluate their calibration. As slide-level classification task, we choose the
prediction of Microsatellite Instability from Colorectal Cancer tissue
sections. We observe that Transformers lead to good results in terms of
classification performance and calibration. When evaluating the classification
performance on a separate dataset, we observe that Transformers generalize
best. The investigation of reliability diagrams provides additional insights to
the Expected Calibration Error metric and we observe that especially
Transformers push the output probabilities to extreme values, which results in
overconfident predictions.
Related papers
- Variational Classification [51.2541371924591]
We derive a variational objective to train the model, analogous to the evidence lower bound (ELBO) used to train variational auto-encoders.
Treating inputs to the softmax layer as samples of a latent variable, our abstracted perspective reveals a potential inconsistency.
We induce a chosen latent distribution, instead of the implicit assumption found in a standard softmax layer.
arXiv Detail & Related papers (2023-05-17T17:47:19Z) - Bayesian Layer Graph Convolutioanl Network for Hyperspetral Image
Classification [24.91896527342631]
Graph convolutional network (GCN) based models have shown impressive performance.
Deep learning frameworks based on point estimation suffer from low generalization and inability to quantify the classification results uncertainty.
In this paper, we propose a Bayesian layer with Bayesian idea as an insertion layer into point estimation based neural networks.
A Generative Adversarial Network (GAN) is built to solve the sample imbalance problem of HSI dataset.
arXiv Detail & Related papers (2022-11-14T12:56:56Z) - On double-descent in uncertainty quantification in overparametrized
models [24.073221004661427]
Uncertainty quantification is a central challenge in reliable and trustworthy machine learning.
We show a trade-off between classification accuracy and calibration, unveiling a double descent like behavior in the calibration curve of optimally regularized estimators.
This is in contrast with the empirical Bayes method, which we show to be well calibrated in our setting despite the higher generalization error and overparametrization.
arXiv Detail & Related papers (2022-10-23T16:01:08Z) - What Makes Graph Neural Networks Miscalibrated? [48.00374886504513]
We conduct a systematic study on the calibration qualities of graph neural networks (GNNs)
We identify five factors which influence the calibration of GNNs: general under-confident tendency, diversity of nodewise predictive distributions, distance to training nodes, relative confidence level, and neighborhood similarity.
We design a novel calibration method named Graph Attention Temperature Scaling (GATS), which is tailored for calibrating graph neural networks.
arXiv Detail & Related papers (2022-10-12T16:41:42Z) - Bayesian Neural Network Versus Ex-Post Calibration For Prediction
Uncertainty [0.2343856409260935]
Probabilistic predictions from neural networks account for predictive uncertainty during classification.
In practice most datasets are trained on non-probabilistic neural networks which by default do not capture this inherent uncertainty.
A plausible alternative to the calibration approach is to use Bayesian neural networks, which directly models a predictive distribution.
arXiv Detail & Related papers (2022-09-29T07:22:19Z) - DOMINO: Domain-aware Model Calibration in Medical Image Segmentation [51.346121016559024]
Modern deep neural networks are poorly calibrated, compromising trustworthiness and reliability.
We propose DOMINO, a domain-aware model calibration method that leverages the semantic confusability and hierarchical similarity between class labels.
Our results show that DOMINO-calibrated deep neural networks outperform non-calibrated models and state-of-the-art morphometric methods in head image segmentation.
arXiv Detail & Related papers (2022-09-13T15:31:52Z) - On Calibration of Graph Neural Networks for Node Classification [29.738179864433445]
Graph neural networks learn entity and edge embeddings for tasks such as node classification and link prediction.
These models achieve good performance with respect to accuracy, but the confidence scores associated with the predictions might not be calibrated.
We propose a topology-aware calibration method that takes the neighboring nodes into account and yields improved calibration.
arXiv Detail & Related papers (2022-06-03T13:48:10Z) - SLA$^2$P: Self-supervised Anomaly Detection with Adversarial
Perturbation [77.71161225100927]
Anomaly detection is a fundamental yet challenging problem in machine learning.
We propose a novel and powerful framework, dubbed as SLA$2$P, for unsupervised anomaly detection.
arXiv Detail & Related papers (2021-11-25T03:53:43Z) - Improved Trainable Calibration Method for Neural Networks on Medical
Imaging Classification [17.941506832422192]
Empirically, neural networks are often miscalibrated and overconfident in their predictions.
We propose a novel calibration approach that maintains the overall classification accuracy while significantly improving model calibration.
arXiv Detail & Related papers (2020-09-09T01:25:53Z) - Unlabelled Data Improves Bayesian Uncertainty Calibration under
Covariate Shift [100.52588638477862]
We develop an approximate Bayesian inference scheme based on posterior regularisation.
We demonstrate the utility of our method in the context of transferring prognostic models of prostate cancer across globally diverse populations.
arXiv Detail & Related papers (2020-06-26T13:50:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.