Difficulty-aware Glaucoma Classification with Multi-Rater Consensus
Modeling
- URL: http://arxiv.org/abs/2007.14848v1
- Date: Wed, 29 Jul 2020 14:04:34 GMT
- Title: Difficulty-aware Glaucoma Classification with Multi-Rater Consensus
Modeling
- Authors: Shuang Yu, Hong-Yu Zhou, Kai Ma, Cheng Bian, Chunyan Chu, Hanruo Liu,
Yefeng Zheng
- Abstract summary: We take advantage of the raw multi-rater gradings to improve the deep learning model performance for the glaucoma classification task.
A multi-branch model structure is proposed to predict the most sensitive, most specifical and a balanced fused result for the input images.
Compared with models trained only with the final ground-truth labels, the proposed method using multi-rater consensus information has achieved superior performance.
- Score: 34.28252351672568
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Medical images are generally labeled by multiple experts before the final
ground-truth labels are determined. Consensus or disagreement among experts
regarding individual images reflects the gradeability and difficulty levels of
the image. However, when being used for model training, only the final
ground-truth label is utilized, while the critical information contained in the
raw multi-rater gradings regarding the image being an easy/hard case is
discarded. In this paper, we aim to take advantage of the raw multi-rater
gradings to improve the deep learning model performance for the glaucoma
classification task. Specifically, a multi-branch model structure is proposed
to predict the most sensitive, most specifical and a balanced fused result for
the input images. In order to encourage the sensitivity branch and specificity
branch to generate consistent results for consensus labels and opposite results
for disagreement labels, a consensus loss is proposed to constrain the output
of the two branches. Meanwhile, the consistency/inconsistency between the
prediction results of the two branches implies the image being an easy/hard
case, which is further utilized to encourage the balanced fusion branch to
concentrate more on the hard cases. Compared with models trained only with the
final ground-truth labels, the proposed method using multi-rater consensus
information has achieved superior performance, and it is also able to estimate
the difficulty levels of individual input images when making the prediction.
Related papers
- Foundation Visual Encoders Are Secretly Few-Shot Anomaly Detectors [58.75916798814376]
We develop a few-shot anomaly detector termed FoundAD.<n>We observe that the anomaly amount in an image directly correlates with the difference in the learnt embeddings.<n>The simple operator acts as an effective tool for anomaly detection to characterize and identify out-of-distribution regions in an image.
arXiv Detail & Related papers (2025-10-02T11:53:20Z) - Decompose-and-Compose: A Compositional Approach to Mitigating Spurious Correlation [2.273629240935727]
We propose Decompose-and-Compose (DaC) to improve correlation shift by combining elements of images.
Based on our observations, models trained with Empirical Risk Minimization (ERM) usually highly attend to either the causal components or the components having a high spurious correlation with the label.
We propose a group-balancing method by intervening on images without requiring group labels or information regarding the spurious features during training.
arXiv Detail & Related papers (2024-02-29T07:24:24Z) - ProbMCL: Simple Probabilistic Contrastive Learning for Multi-label Visual Classification [16.415582577355536]
Multi-label image classification presents a challenging task in many domains, including computer vision and medical imaging.
Recent advancements have introduced graph-based and transformer-based methods to improve performance and capture label dependencies.
We propose Probabilistic Multi-label Contrastive Learning (ProbMCL), a novel framework to address these challenges.
arXiv Detail & Related papers (2024-01-02T22:15:20Z) - Simple and Efficient Confidence Score for Grading Whole Slide Images [0.7349727826230862]
We propose a new score to measure the confidence of AI models in grading tasks.
Our confidence score is specifically adapted to ordinal output variables, is versatile and does not require extra training or additional inferences.
We show that the score is capable of accurately identifying mispredicted slides and that accuracy for high confidence decisions is significantly higher than for low-confidence decisions.
arXiv Detail & Related papers (2023-03-08T14:15:43Z) - Modulated Contrast for Versatile Image Synthesis [60.304183493234376]
MoNCE is a versatile metric that introduces image contrast to learn a calibrated metric for the perception of multifaceted inter-image distances.
We introduce optimal transport in MoNCE to modulate the pushing force of negative samples collaboratively across multiple contrastive objectives.
arXiv Detail & Related papers (2022-03-17T14:03:46Z) - Label fusion and training methods for reliable representation of
inter-rater uncertainty [0.0]
Training deep learning networks with annotations from multiple raters mitigates the model's bias towards a single expert.
Various methods exist to take into account different expert labels.
We compare three label fusion methods: STAPLE, average of the rater's segmentation, and random sampling each rater's segmentation during training.
arXiv Detail & Related papers (2022-02-15T16:35:47Z) - Palette: Image-to-Image Diffusion Models [50.268441533631176]
We introduce Palette, a simple and general framework for image-to-image translation using conditional diffusion models.
On four challenging image-to-image translation tasks, Palette outperforms strong GAN and regression baselines.
We report several sample quality scores including FID, Inception Score, Classification Accuracy of a pre-trained ResNet-50, and Perceptual Distance against reference images.
arXiv Detail & Related papers (2021-11-10T17:49:29Z) - Multi-Label Image Classification with Contrastive Learning [57.47567461616912]
We show that a direct application of contrastive learning can hardly improve in multi-label cases.
We propose a novel framework for multi-label classification with contrastive learning in a fully supervised setting.
arXiv Detail & Related papers (2021-07-24T15:00:47Z) - Self-Ensembling Contrastive Learning for Semi-Supervised Medical Image
Segmentation [6.889911520730388]
We aim to boost the performance of semi-supervised learning for medical image segmentation with limited labels.
We learn latent representations directly at feature-level by imposing contrastive loss on unlabeled images.
We conduct experiments on an MRI and a CT segmentation dataset and demonstrate that the proposed method achieves state-of-the-art performance.
arXiv Detail & Related papers (2021-05-27T03:27:58Z) - Delving into Inter-Image Invariance for Unsupervised Visual
Representations [108.33534231219464]
We present a study to better understand the role of inter-image invariance learning.
Online labels converge faster than offline labels.
Semi-hard negative samples are more reliable and unbiased than hard negative samples.
arXiv Detail & Related papers (2020-08-26T17:44:23Z) - Semi-supervised Medical Image Classification with Relation-driven
Self-ensembling Model [71.80319052891817]
We present a relation-driven semi-supervised framework for medical image classification.
It exploits the unlabeled data by encouraging the prediction consistency of given input under perturbations.
Our method outperforms many state-of-the-art semi-supervised learning methods on both single-label and multi-label image classification scenarios.
arXiv Detail & Related papers (2020-05-15T06:57:54Z) - I Am Going MAD: Maximum Discrepancy Competition for Comparing
Classifiers Adaptively [135.7695909882746]
We name the MAximum Discrepancy (MAD) competition.
We adaptively sample a small test set from an arbitrarily large corpus of unlabeled images.
Human labeling on the resulting model-dependent image sets reveals the relative performance of the competing classifiers.
arXiv Detail & Related papers (2020-02-25T03:32:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.