DOMINO: Domain-aware Model Calibration in Medical Image Segmentation
- URL: http://arxiv.org/abs/2209.06077v1
- Date: Tue, 13 Sep 2022 15:31:52 GMT
- Title: DOMINO: Domain-aware Model Calibration in Medical Image Segmentation
- Authors: Skylar E. Stolte, Kyle Volle, Aprinda Indahlastari, Alejandro Albizu,
Adam J. Woods, Kevin Brink, Matthew Hale, Ruogu Fang
- Abstract summary: Modern deep neural networks are poorly calibrated, compromising trustworthiness and reliability.
We propose DOMINO, a domain-aware model calibration method that leverages the semantic confusability and hierarchical similarity between class labels.
Our results show that DOMINO-calibrated deep neural networks outperform non-calibrated models and state-of-the-art morphometric methods in head image segmentation.
- Score: 51.346121016559024
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Model calibration measures the agreement between the predicted probability
estimates and the true correctness likelihood. Proper model calibration is
vital for high-risk applications. Unfortunately, modern deep neural networks
are poorly calibrated, compromising trustworthiness and reliability. Medical
image segmentation particularly suffers from this due to the natural
uncertainty of tissue boundaries. This is exasperated by their loss functions,
which favor overconfidence in the majority classes. We address these challenges
with DOMINO, a domain-aware model calibration method that leverages the
semantic confusability and hierarchical similarity between class labels. Our
experiments demonstrate that our DOMINO-calibrated deep neural networks
outperform non-calibrated models and state-of-the-art morphometric methods in
head image segmentation. Our results show that our method can consistently
achieve better calibration, higher accuracy, and faster inference times than
these methods, especially on rarer classes. This performance is attributed to
our domain-aware regularization to inform semantic model calibration. These
findings show the importance of semantic ties between class labels in building
confidence in deep learning models. The framework has the potential to improve
the trustworthiness and reliability of generic medical image segmentation
models. The code for this article is available at:
https://github.com/lab-smile/DOMINO.
Related papers
- Multiclass Alignment of Confidence and Certainty for Network Calibration [10.15706847741555]
Recent studies reveal that deep neural networks (DNNs) are prone to making overconfident predictions.
We propose a new train-time calibration method, which features a simple, plug-and-play auxiliary loss known as multi-class alignment of predictive mean confidence and predictive certainty (MACC)
Our method achieves state-of-the-art calibration performance for both in-domain and out-domain predictions.
arXiv Detail & Related papers (2023-09-06T00:56:24Z) - Bridging Precision and Confidence: A Train-Time Loss for Calibrating
Object Detection [58.789823426981044]
We propose a novel auxiliary loss formulation that aims to align the class confidence of bounding boxes with the accurateness of predictions.
Our results reveal that our train-time loss surpasses strong calibration baselines in reducing calibration error for both in and out-domain scenarios.
arXiv Detail & Related papers (2023-03-25T08:56:21Z) - Multi-Head Multi-Loss Model Calibration [13.841172927454204]
We introduce a form of simplified ensembling that bypasses the costly training and inference of deep ensembles.
Specifically, each head is trained to minimize a weighted Cross-Entropy loss, but the weights are different among the different branches.
We show that the resulting averaged predictions can achieve excellent calibration without sacrificing accuracy in two challenging datasets.
arXiv Detail & Related papers (2023-03-02T09:32:32Z) - DOMINO: Domain-aware Loss for Deep Learning Calibration [49.485186880996125]
This paper proposes a novel domain-aware loss function to calibrate deep learning models.
The proposed loss function applies a class-wise penalty based on the similarity between classes within a given target domain.
arXiv Detail & Related papers (2023-02-10T09:47:46Z) - On Calibrating Semantic Segmentation Models: Analyses and An Algorithm [51.85289816613351]
We study the problem of semantic segmentation calibration.
Model capacity, crop size, multi-scale testing, and prediction correctness have impact on calibration.
We propose a simple, unifying, and effective approach, namely selective scaling.
arXiv Detail & Related papers (2022-12-22T22:05:16Z) - Improved post-hoc probability calibration for out-of-domain MRI
segmentation [18.089067656236125]
Well-calibrated probabilities allow radiologists to identify regions where model-predicted segmentations are unreliable.
These unreliable predictions often occur to out-of-domain (OOD) images that are caused by imaging artifacts or unseen imaging protocols.
We propose a novel post-hoc calibration model to reduce the calibration error when confronted with OOD images.
arXiv Detail & Related papers (2022-08-04T20:13:19Z) - On the Dark Side of Calibration for Modern Neural Networks [65.83956184145477]
We show the breakdown of expected calibration error (ECE) into predicted confidence and refinement.
We highlight that regularisation based calibration only focuses on naively reducing a model's confidence.
We find that many calibration approaches with the likes of label smoothing, mixup etc. lower the utility of a DNN by degrading its refinement.
arXiv Detail & Related papers (2021-06-17T11:04:14Z) - Improved Trainable Calibration Method for Neural Networks on Medical
Imaging Classification [17.941506832422192]
Empirically, neural networks are often miscalibrated and overconfident in their predictions.
We propose a novel calibration approach that maintains the overall classification accuracy while significantly improving model calibration.
arXiv Detail & Related papers (2020-09-09T01:25:53Z) - On Calibration of Mixup Training for Deep Neural Networks [1.6242924916178283]
We argue and provide empirical evidence that, due to its fundamentals, Mixup does not necessarily improve calibration.
Our loss is inspired by Bayes decision theory and introduces a new training framework for designing losses for probabilistic modelling.
We provide state-of-the-art accuracy with consistent improvements in calibration performance.
arXiv Detail & Related papers (2020-03-22T16:54:31Z) - Calibrating Deep Neural Networks using Focal Loss [77.92765139898906]
Miscalibration is a mismatch between a model's confidence and its correctness.
We show that focal loss allows us to learn models that are already very well calibrated.
We show that our approach achieves state-of-the-art calibration without compromising on accuracy in almost all cases.
arXiv Detail & Related papers (2020-02-21T17:35:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.