D-LEMA: Deep Learning Ensembles from Multiple Annotations -- Application
to Skin Lesion Segmentation
- URL: http://arxiv.org/abs/2012.07206v2
- Date: Thu, 29 Apr 2021 01:31:40 GMT
- Title: D-LEMA: Deep Learning Ensembles from Multiple Annotations -- Application
to Skin Lesion Segmentation
- Authors: Zahra Mirikharaji, Kumar Abhishek, Saeed Izadi, Ghassan Hamarneh
- Abstract summary: Leveraging a collection of annotators' opinions for an image is an interesting way of estimating a gold standard.
We propose an approach to handle annotators' disagreements when training a deep model.
- Score: 14.266037264648533
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Medical image segmentation annotations suffer from inter- and intra-observer
variations even among experts due to intrinsic differences in human annotators
and ambiguous boundaries. Leveraging a collection of annotators' opinions for
an image is an interesting way of estimating a gold standard. Although training
deep models in a supervised setting with a single annotation per image has been
extensively studied, generalizing their training to work with datasets
containing multiple annotations per image remains a fairly unexplored problem.
In this paper, we propose an approach to handle annotators' disagreements when
training a deep model. To this end, we propose an ensemble of Bayesian fully
convolutional networks (FCNs) for the segmentation task by considering two
major factors in the aggregation of multiple ground truth annotations: (1)
handling contradictory annotations in the training data originating from
inter-annotator disagreements and (2) improving confidence calibration through
the fusion of base models' predictions. We demonstrate the superior performance
of our approach on the ISIC Archive and explore the generalization performance
of our proposed method by cross-dataset evaluation on the PH2 and DermoFit
datasets.
Related papers
- Multi-rater Prompting for Ambiguous Medical Image Segmentation [12.452584289825849]
Multi-rater annotations commonly occur when medical images are independently annotated by multiple experts (raters)
We propose a multi-rater prompt-based approach to address these two challenges altogether.
arXiv Detail & Related papers (2024-04-11T09:13:50Z) - Robust Training of Federated Models with Extremely Label Deficiency [84.00832527512148]
Federated semi-supervised learning (FSSL) has emerged as a powerful paradigm for collaboratively training machine learning models using distributed data with label deficiency.
We propose a novel twin-model paradigm, called Twin-sight, designed to enhance mutual guidance by providing insights from different perspectives of labeled and unlabeled data.
Our comprehensive experiments on four benchmark datasets provide substantial evidence that Twin-sight can significantly outperform state-of-the-art methods across various experimental settings.
arXiv Detail & Related papers (2024-02-22T10:19:34Z) - MSE-Nets: Multi-annotated Semi-supervised Ensemble Networks for
Improving Segmentation of Medical Image with Ambiguous Boundaries [21.513613620213754]
We propose Multi-annotated Semi-supervised Ensemble Networks (MSE-Nets) for learning medical image segmentation from limited multi-annotated data.
We introduce the Network Pairwise Consistency Enhancement (NPCE) module and Multi-Network PseudoSupervised (MNPS) module to enhance MSE-Nets for the segmentation task.
Experiments on the ISIC dataset show that we reduced the demand for multi-annotated data by 97.75% and narrowed the gap with the best fully-supervised baseline to just a Jaccard index of 4%.
arXiv Detail & Related papers (2023-11-17T08:14:24Z) - Consistency Regularisation in Varying Contexts and Feature Perturbations
for Semi-Supervised Semantic Segmentation of Histology Images [14.005379068469361]
We present a consistency based semi-supervised learning (SSL) approach that can help mitigate this challenge.
SSL models might also be susceptible to changing context and features perturbations exhibiting poor generalisation due to the limited training data.
We show that cross-consistency training makes the encoder features invariant to different perturbations and improves the prediction confidence.
arXiv Detail & Related papers (2023-01-30T18:21:57Z) - Phoneme Segmentation Using Self-Supervised Speech Models [13.956691231452336]
We apply transfer learning to the task of phoneme segmentation and demonstrate the utility of representations learned in self-supervised pre-training for the task.
Our model extends transformer-style encoders with strategically placed convolutions that manipulate features learned in pre-training.
arXiv Detail & Related papers (2022-11-02T19:57:31Z) - Semantics-Depth-Symbiosis: Deeply Coupled Semi-Supervised Learning of
Semantics and Depth [83.94528876742096]
We tackle the MTL problem of two dense tasks, ie, semantic segmentation and depth estimation, and present a novel attention module called Cross-Channel Attention Module (CCAM)
In a true symbiotic spirit, we then formulate a novel data augmentation for the semantic segmentation task using predicted depth called AffineMix, and a simple depth augmentation using predicted semantics called ColorAug.
Finally, we validate the performance gain of the proposed method on the Cityscapes dataset, which helps us achieve state-of-the-art results for a semi-supervised joint model based on depth and semantic
arXiv Detail & Related papers (2022-06-21T17:40:55Z) - Dense Contrastive Visual-Linguistic Pretraining [53.61233531733243]
Several multimodal representation learning approaches have been proposed that jointly represent image and text.
These approaches achieve superior performance by capturing high-level semantic information from large-scale multimodal pretraining.
We propose unbiased Dense Contrastive Visual-Linguistic Pretraining to replace the region regression and classification with cross-modality region contrastive learning.
arXiv Detail & Related papers (2021-09-24T07:20:13Z) - SOSD-Net: Joint Semantic Object Segmentation and Depth Estimation from
Monocular images [94.36401543589523]
We introduce the concept of semantic objectness to exploit the geometric relationship of these two tasks.
We then propose a Semantic Object and Depth Estimation Network (SOSD-Net) based on the objectness assumption.
To the best of our knowledge, SOSD-Net is the first network that exploits the geometry constraint for simultaneous monocular depth estimation and semantic segmentation.
arXiv Detail & Related papers (2021-01-19T02:41:03Z) - Gradient-Induced Co-Saliency Detection [81.54194063218216]
Co-saliency detection (Co-SOD) aims to segment the common salient foreground in a group of relevant images.
In this paper, inspired by human behavior, we propose a gradient-induced co-saliency detection method.
arXiv Detail & Related papers (2020-04-28T08:40:55Z) - Deep Semantic Matching with Foreground Detection and Cycle-Consistency [103.22976097225457]
We address weakly supervised semantic matching based on a deep network.
We explicitly estimate the foreground regions to suppress the effect of background clutter.
We develop cycle-consistent losses to enforce the predicted transformations across multiple images to be geometrically plausible and consistent.
arXiv Detail & Related papers (2020-03-31T22:38:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.