Efficient Medical Image Assessment via Self-supervised Learning
- URL: http://arxiv.org/abs/2209.14434v1
- Date: Wed, 28 Sep 2022 21:39:00 GMT
- Title: Efficient Medical Image Assessment via Self-supervised Learning
- Authors: Chun-Yin Huang, Qi Lei, and Xiaoxiao Li
- Abstract summary: High-performance deep learning methods typically rely on large annotated training datasets.
We propose a novel and efficient data assessment strategy to rank the quality of unlabeled medical image data.
Motivated by theoretical implication of SSL embedding space, we leverage a Masked Autoencoder for feature extraction.
- Score: 27.969767956918503
- License: http://creativecommons.org/publicdomain/zero/1.0/
- Abstract: High-performance deep learning methods typically rely on large annotated
training datasets, which are difficult to obtain in many clinical applications
due to the high cost of medical image labeling. Existing data assessment
methods commonly require knowing the labels in advance, which are not feasible
to achieve our goal of 'knowing which data to label.' To this end, we formulate
and propose a novel and efficient data assessment strategy, EXponentiAl
Marginal sINgular valuE (EXAMINE) score, to rank the quality of unlabeled
medical image data based on their useful latent representations extracted via
Self-supervised Learning (SSL) networks. Motivated by theoretical implication
of SSL embedding space, we leverage a Masked Autoencoder for feature
extraction. Furthermore, we evaluate data quality based on the marginal change
of the largest singular value after excluding the data point in the dataset. We
conduct extensive experiments on a pathology dataset. Our results indicate the
effectiveness and efficiency of our proposed methods for selecting the most
valuable data to label.
Related papers
- Integration of Self-Supervised BYOL in Semi-Supervised Medical Image Recognition [10.317372960942972]
We propose an innovative approach by integrating self-supervised learning into semi-supervised models to enhance medical image recognition.
Our approach optimally leverages unlabeled data, outperforming existing methods in terms of accuracy for medical image recognition.
arXiv Detail & Related papers (2024-04-16T09:12:16Z) - Pseudo Label-Guided Data Fusion and Output Consistency for
Semi-Supervised Medical Image Segmentation [9.93871075239635]
We propose the PLGDF framework, which builds upon the mean teacher network for segmenting medical images with less annotation.
We propose a novel pseudo-label utilization scheme, which combines labeled and unlabeled data to augment the dataset effectively.
Our framework yields superior performance compared to six state-of-the-art semi-supervised learning methods.
arXiv Detail & Related papers (2023-11-17T06:36:43Z) - Clinically Acceptable Segmentation of Organs at Risk in Cervical Cancer
Radiation Treatment from Clinically Available Annotations [0.0]
We present an approach to learn a deep learning model for the automatic segmentation of Organs at Risk (OARs) in cervical cancer radiation treatment.
We employ simples for automatic data cleaning to minimize data inhomogeneity, label noise, and missing annotations.
We develop a semi-supervised learning approach utilizing a teacher-student setup, annotation imputation, and uncertainty-guided training to learn in presence of missing annotations.
arXiv Detail & Related papers (2023-02-21T13:24:40Z) - Temporal Output Discrepancy for Loss Estimation-based Active Learning [65.93767110342502]
We present a novel deep active learning approach that queries the oracle for data annotation when the unlabeled sample is believed to incorporate high loss.
Our approach achieves superior performances than the state-of-the-art active learning methods on image classification and semantic segmentation tasks.
arXiv Detail & Related papers (2022-12-20T19:29:37Z) - Cut-Paste Consistency Learning for Semi-Supervised Lesion Segmentation [0.20305676256390934]
Semi-supervised learning has the potential to improve the data-efficiency of training data-hungry deep neural networks.
We present a simple semi-supervised learning method for lesion segmentation tasks based on the ideas of cut-paste augmentation and consistency regularization.
arXiv Detail & Related papers (2022-10-01T04:43:54Z) - Self-Supervised Learning as a Means To Reduce the Need for Labeled Data
in Medical Image Analysis [64.4093648042484]
We use a dataset of chest X-ray images with bounding box labels for 13 different classes of anomalies.
We show that it is possible to achieve similar performance to a fully supervised model in terms of mean average precision and accuracy with only 60% of the labeled data.
arXiv Detail & Related papers (2022-06-01T09:20:30Z) - A Real Use Case of Semi-Supervised Learning for Mammogram Classification
in a Local Clinic of Costa Rica [0.5541644538483946]
Training a deep learning model requires a considerable amount of labeled images.
A number of publicly available datasets have been built with data from different hospitals and clinics.
The use of the semi-supervised deep learning approach known as MixMatch, to leverage the usage of unlabeled data is proposed and evaluated.
arXiv Detail & Related papers (2021-07-24T22:26:50Z) - Active learning for medical code assignment [55.99831806138029]
We demonstrate the effectiveness of Active Learning (AL) in multi-label text classification in the clinical domain.
We apply a set of well-known AL methods to help automatically assign ICD-9 codes on the MIMIC-III dataset.
Our results show that the selection of informative instances provides satisfactory classification with a significantly reduced training set.
arXiv Detail & Related papers (2021-04-12T18:11:17Z) - Uncovering the structure of clinical EEG signals with self-supervised
learning [64.4754948595556]
Supervised learning paradigms are often limited by the amount of labeled data that is available.
This phenomenon is particularly problematic in clinically-relevant data, such as electroencephalography (EEG)
By extracting information from unlabeled data, it might be possible to reach competitive performance with deep neural networks.
arXiv Detail & Related papers (2020-07-31T14:34:47Z) - Omni-supervised Facial Expression Recognition via Distilled Data [120.11782405714234]
We propose omni-supervised learning to exploit reliable samples in a large amount of unlabeled data for network training.
We experimentally verify that the new dataset can significantly improve the ability of the learned FER model.
To tackle this, we propose to apply a dataset distillation strategy to compress the created dataset into several informative class-wise images.
arXiv Detail & Related papers (2020-05-18T09:36:51Z) - Semi-supervised Medical Image Classification with Relation-driven
Self-ensembling Model [71.80319052891817]
We present a relation-driven semi-supervised framework for medical image classification.
It exploits the unlabeled data by encouraging the prediction consistency of given input under perturbations.
Our method outperforms many state-of-the-art semi-supervised learning methods on both single-label and multi-label image classification scenarios.
arXiv Detail & Related papers (2020-05-15T06:57:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.