Improving Medical Annotation Quality to Decrease Labeling Burden Using
Stratified Noisy Cross-Validation
- URL: http://arxiv.org/abs/2009.10858v1
- Date: Tue, 22 Sep 2020 23:32:59 GMT
- Title: Improving Medical Annotation Quality to Decrease Labeling Burden Using
Stratified Noisy Cross-Validation
- Authors: Joy Hsu, Sonia Phene, Akinori Mitani, Jieying Luo, Naama Hammel,
Jonathan Krause, Rory Sayres
- Abstract summary: Variability in diagnosis of medical images is well established; variability in training and attention to task among medical labelers may exacerbate this issue.
Noisy Cross-Validation splits the training data into halves, and has been shown to identify low-quality labels in computer vision tasks.
In this work we introduce Stratified Noisy Cross-Validation (SNCV), an extension of noisy cross validation.
- Score: 3.690031561736533
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: As machine learning has become increasingly applied to medical imaging data,
noise in training labels has emerged as an important challenge. Variability in
diagnosis of medical images is well established; in addition, variability in
training and attention to task among medical labelers may exacerbate this
issue. Methods for identifying and mitigating the impact of low quality labels
have been studied, but are not well characterized in medical imaging tasks. For
instance, Noisy Cross-Validation splits the training data into halves, and has
been shown to identify low-quality labels in computer vision tasks; but it has
not been applied to medical imaging tasks specifically. In this work we
introduce Stratified Noisy Cross-Validation (SNCV), an extension of noisy cross
validation. SNCV can provide estimates of confidence in model predictions by
assigning a quality score to each example; stratify labels to handle class
imbalance; and identify likely low-quality labels to analyze the causes. We
assess performance of SNCV on diagnosis of glaucoma suspect risk from retinal
fundus photographs, a clinically important yet nuanced labeling task. Using
training data from a previously-published deep learning model, we compute a
continuous quality score (QS) for each training example. We relabel 1,277
low-QS examples using a trained glaucoma specialist; the new labels agree with
the SNCV prediction over the initial label >85% of the time, indicating that
low-QS examples mostly reflect labeler errors. We then quantify the impact of
training with only high-QS labels, showing that strong model performance may be
obtained with many fewer examples. By applying the method to randomly
sub-sampled training dataset, we show that our method can reduce labelling
burden by approximately 50% while achieving model performance non-inferior to
using the full dataset on multiple held-out test sets.
Related papers
- Self-Supervised Learning as a Means To Reduce the Need for Labeled Data
in Medical Image Analysis [64.4093648042484]
We use a dataset of chest X-ray images with bounding box labels for 13 different classes of anomalies.
We show that it is possible to achieve similar performance to a fully supervised model in terms of mean average precision and accuracy with only 60% of the labeled data.
arXiv Detail & Related papers (2022-06-01T09:20:30Z) - Robust Medical Image Classification from Noisy Labeled Data with Global
and Local Representation Guided Co-training [73.60883490436956]
We propose a novel collaborative training paradigm with global and local representation learning for robust medical image classification.
We employ the self-ensemble model with a noisy label filter to efficiently select the clean and noisy samples.
We also design a novel global and local representation learning scheme to implicitly regularize the networks to utilize noisy samples.
arXiv Detail & Related papers (2022-05-10T07:50:08Z) - Incorporating Semi-Supervised and Positive-Unlabeled Learning for
Boosting Full Reference Image Quality Assessment [73.61888777504377]
Full-reference (FR) image quality assessment (IQA) evaluates the visual quality of a distorted image by measuring its perceptual difference with pristine-quality reference.
Unlabeled data can be easily collected from an image degradation or restoration process, making it encouraging to exploit unlabeled training data to boost FR-IQA performance.
In this paper, we suggest to incorporate semi-supervised and positive-unlabeled (PU) learning for exploiting unlabeled data while mitigating the adverse effect of outliers.
arXiv Detail & Related papers (2022-04-19T09:10:06Z) - Improving Medical Image Classification with Label Noise Using
Dual-uncertainty Estimation [72.0276067144762]
We discuss and define the two common types of label noise in medical images.
We propose an uncertainty estimation-based framework to handle these two label noise amid the medical image classification task.
arXiv Detail & Related papers (2021-02-28T14:56:45Z) - Learning Image Labels On-the-fly for Training Robust Classification
Models [13.669654965671604]
We show how noisy annotations (e.g., from different algorithm-based labelers) can be utilized together and mutually benefit the learning of classification tasks.
A meta-training based label-sampling module is designed to attend the labels that benefit the model learning the most through additional back-propagation processes.
arXiv Detail & Related papers (2020-09-22T05:38:44Z) - Multi-label Thoracic Disease Image Classification with Cross-Attention
Networks [65.37531731899837]
We propose a novel scheme of Cross-Attention Networks (CAN) for automated thoracic disease classification from chest x-ray images.
We also design a new loss function that beyond cross-entropy loss to help cross-attention process and is able to overcome the imbalance between classes and easy-dominated samples within each class.
arXiv Detail & Related papers (2020-07-21T14:37:00Z) - Semi-supervised Medical Image Classification with Relation-driven
Self-ensembling Model [71.80319052891817]
We present a relation-driven semi-supervised framework for medical image classification.
It exploits the unlabeled data by encouraging the prediction consistency of given input under perturbations.
Our method outperforms many state-of-the-art semi-supervised learning methods on both single-label and multi-label image classification scenarios.
arXiv Detail & Related papers (2020-05-15T06:57:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.