Improving Medical Image Classification in Noisy Labels Using Only
Self-supervised Pretraining
- URL: http://arxiv.org/abs/2308.04551v1
- Date: Tue, 8 Aug 2023 19:45:06 GMT
- Title: Improving Medical Image Classification in Noisy Labels Using Only
Self-supervised Pretraining
- Authors: Bidur Khanal, Binod Bhattarai, Bishesh Khanal, Cristian A. Linte
- Abstract summary: Noisy labels hurt deep learning-based supervised image classification performance as the models may overfit the noise and learn corrupted feature extractors.
In this work, we explore contrastive and pretext task-based self-supervised pretraining to initialize the weights of a deep learning classification model for two medical datasets with self-induced noisy labels.
Our results show that models with pretrained weights obtained from self-supervised learning can effectively learn better features and improve robustness against noisy labels.
- Score: 9.01547574908261
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Noisy labels hurt deep learning-based supervised image classification
performance as the models may overfit the noise and learn corrupted feature
extractors. For natural image classification training with noisy labeled data,
model initialization with contrastive self-supervised pretrained weights has
shown to reduce feature corruption and improve classification performance.
However, no works have explored: i) how other self-supervised approaches, such
as pretext task-based pretraining, impact the learning with noisy label, and
ii) any self-supervised pretraining methods alone for medical images in noisy
label settings. Medical images often feature smaller datasets and subtle inter
class variations, requiring human expertise to ensure correct classification.
Thus, it is not clear if the methods improving learning with noisy labels in
natural image datasets such as CIFAR would also help with medical images. In
this work, we explore contrastive and pretext task-based self-supervised
pretraining to initialize the weights of a deep learning classification model
for two medical datasets with self-induced noisy labels -- NCT-CRC-HE-100K
tissue histological images and COVID-QU-Ex chest X-ray images. Our results show
that models initialized with pretrained weights obtained from self-supervised
learning can effectively learn better features and improve robustness against
noisy labels.
Related papers
- Label-noise-tolerant medical image classification via self-attention and
self-supervised learning [5.6827706625306345]
We propose a noise-robust training approach to mitigate the adverse effects of noisy labels in medical image classification.
Specifically, we incorporate contrastive learning and intra-group attention mixup strategies into the vanilla supervised learning.
Rigorous experiments validate that our noise-robust method with contrastive learning and attention mixup can effectively handle with label noise.
arXiv Detail & Related papers (2023-06-16T09:37:16Z) - Seamless Iterative Semi-Supervised Correction of Imperfect Labels in
Microscopy Images [57.42492501915773]
In-vitro tests are an alternative to animal testing for the toxicity of medical devices.
Human fatigue plays a role in error making, making the use of deep learning appealing.
We propose Seamless Iterative Semi-Supervised correction of Imperfect labels (SISSI)
Our method successfully provides an adaptive early learning correction technique for object detection.
arXiv Detail & Related papers (2022-08-05T18:52:20Z) - Self-Supervised Learning as a Means To Reduce the Need for Labeled Data
in Medical Image Analysis [64.4093648042484]
We use a dataset of chest X-ray images with bounding box labels for 13 different classes of anomalies.
We show that it is possible to achieve similar performance to a fully supervised model in terms of mean average precision and accuracy with only 60% of the labeled data.
arXiv Detail & Related papers (2022-06-01T09:20:30Z) - Deep Image Retrieval is not Robust to Label Noise [0.0]
We show that image retrieval methods are less robust to label noise than image classification ones.
For the first time, we investigate different types of label noise specific to image retrieval tasks.
arXiv Detail & Related papers (2022-05-23T11:04:09Z) - Robust Medical Image Classification from Noisy Labeled Data with Global
and Local Representation Guided Co-training [73.60883490436956]
We propose a novel collaborative training paradigm with global and local representation learning for robust medical image classification.
We employ the self-ensemble model with a noisy label filter to efficiently select the clean and noisy samples.
We also design a novel global and local representation learning scheme to implicitly regularize the networks to utilize noisy samples.
arXiv Detail & Related papers (2022-05-10T07:50:08Z) - Self-Paced Contrastive Learning for Semi-supervisedMedical Image
Segmentation with Meta-labels [6.349708371894538]
We propose to adapt contrastive learning to work with meta-label annotations.
We use the meta-labels for pre-training the image encoder as well as to regularize a semi-supervised training.
Results on three different medical image segmentation datasets show that our approach highly boosts the performance of a model trained on a few scans.
arXiv Detail & Related papers (2021-07-29T04:30:46Z) - Noisy Label Learning for Large-scale Medical Image Classification [37.79118840129632]
We adapt a state-of-the-art noisy-label multi-class training approach to learn a multi-label classifier for the dataset Chest X-ray14.
We show that the majority of label noise on Chest X-ray14 is present in the class 'No Finding', which is intuitively correct because this is the most likely class to contain one or more of the 14 diseases due to labelling mistakes.
arXiv Detail & Related papers (2021-03-06T07:42:36Z) - Improving Medical Image Classification with Label Noise Using
Dual-uncertainty Estimation [72.0276067144762]
We discuss and define the two common types of label noise in medical images.
We propose an uncertainty estimation-based framework to handle these two label noise amid the medical image classification task.
arXiv Detail & Related papers (2021-02-28T14:56:45Z) - Annotation-Efficient Learning for Medical Image Segmentation based on
Noisy Pseudo Labels and Adversarial Learning [12.781598229608983]
We propose an annotation-efficient learning framework for medical image segmentation.
We use an improved Cycle-Consistent Generative Adversarial Network (GAN) to learn from a set of unpaired medical images and auxiliary masks.
We validated our framework with two situations: objects with a simple shape model like optic disc in fundus images and fetal head in ultrasound images, and complex structures like lung in X-Ray images and liver in CT images.
arXiv Detail & Related papers (2020-12-29T03:22:41Z) - Attention-Aware Noisy Label Learning for Image Classification [97.26664962498887]
Deep convolutional neural networks (CNNs) learned on large-scale labeled samples have achieved remarkable progress in computer vision.
The cheapest way to obtain a large body of labeled visual data is to crawl from websites with user-supplied labels, such as Flickr.
This paper proposes the attention-aware noisy label learning approach to improve the discriminative capability of the network trained on datasets with potential label noise.
arXiv Detail & Related papers (2020-09-30T15:45:36Z) - Multi-label Thoracic Disease Image Classification with Cross-Attention
Networks [65.37531731899837]
We propose a novel scheme of Cross-Attention Networks (CAN) for automated thoracic disease classification from chest x-ray images.
We also design a new loss function that beyond cross-entropy loss to help cross-attention process and is able to overcome the imbalance between classes and easy-dominated samples within each class.
arXiv Detail & Related papers (2020-07-21T14:37:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.