Tackling the Incomplete Annotation Issue in Universal Lesion Detection
Task By Exploratory Training
- URL: http://arxiv.org/abs/2309.13306v1
- Date: Sat, 23 Sep 2023 08:44:07 GMT
- Title: Tackling the Incomplete Annotation Issue in Universal Lesion Detection
Task By Exploratory Training
- Authors: Xiaoyu Bai, Benteng Ma, Changyang Li and Yong Xia
- Abstract summary: Universal lesion detection has great value for clinical practice as it aims to detect lesions in multiple organs on medical images.
Deep learning methods have shown promising results, but demanding large volumes of annotated data for training.
We introduce a teacher-student detection model as basis, where the teacher's predictions are combined with incomplete annotations to train the student.
- Score: 10.627977735890191
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Universal lesion detection has great value for clinical practice as it aims
to detect various types of lesions in multiple organs on medical images. Deep
learning methods have shown promising results, but demanding large volumes of
annotated data for training. However, annotating medical images is costly and
requires specialized knowledge. The diverse forms and contrasts of objects in
medical images make fully annotation even more challenging, resulting in
incomplete annotations. Directly training ULD detectors on such datasets can
yield suboptimal results. Pseudo-label-based methods examine the training data
and mine unlabelled objects for retraining, which have shown to be effective to
tackle this issue. Presently, top-performing methods rely on a dynamic
label-mining mechanism, operating at the mini-batch level. However, the model's
performance varies at different iterations, leading to inconsistencies in the
quality of the mined labels and limits their performance enhancement. Inspired
by the observation that deep models learn concepts with increasing complexity,
we introduce an innovative exploratory training to assess the reliability of
mined lesions over time. Specifically, we introduce a teacher-student detection
model as basis, where the teacher's predictions are combined with incomplete
annotations to train the student. Additionally, we design a prediction bank to
record high-confidence predictions. Each sample is trained several times,
allowing us to get a sequence of records for each sample. If a prediction
consistently appears in the record sequence, it is likely to be a true object,
otherwise it may just a noise. This serves as a crucial criterion for selecting
reliable mined lesions for retraining. Our experimental results substantiate
that the proposed framework surpasses state-of-the-art methods on two medical
image datasets, demonstrating its superior performance.
Related papers
- A Comprehensive Library for Benchmarking Multi-class Visual Anomaly Detection [52.228708947607636]
This paper introduces a comprehensive visual anomaly detection benchmark, ADer, which is a modular framework for new methods.
The benchmark includes multiple datasets from industrial and medical domains, implementing fifteen state-of-the-art methods and nine comprehensive metrics.
We objectively reveal the strengths and weaknesses of different methods and provide insights into the challenges and future directions of multi-class visual anomaly detection.
arXiv Detail & Related papers (2024-06-05T13:40:07Z) - Boosting Few-Shot Learning with Disentangled Self-Supervised Learning and Meta-Learning for Medical Image Classification [8.975676404678374]
We present a strategy for improving the performance and generalization capabilities of models trained in low-data regimes.
The proposed method starts with a pre-training phase, where features learned in a self-supervised learning setting are disentangled to improve the robustness of the representations for downstream tasks.
We then introduce a meta-fine-tuning step, leveraging related classes between meta-training and meta-testing phases but varying the level.
arXiv Detail & Related papers (2024-03-26T09:36:20Z) - Self-Supervised Pre-Training with Contrastive and Masked Autoencoder
Methods for Dealing with Small Datasets in Deep Learning for Medical Imaging [8.34398674359296]
Deep learning in medical imaging has the potential to minimize the risk of diagnostic errors, reduce radiologist workload, and accelerate diagnosis.
Training such deep learning models requires large and accurate datasets, with annotations for all training samples.
To address this challenge, deep learning models can be pre-trained on large image datasets without annotations using methods from the field of self-supervised learning.
arXiv Detail & Related papers (2023-08-12T11:31:01Z) - An End-to-End Framework For Universal Lesion Detection With Missing
Annotations [24.902835211573628]
We present a novel end-to-end framework for mining unlabeled lesions while simultaneously training the detector.
Our framework follows the teacher-student paradigm. High-confidence predictions are combined with partially-labeled ground truth for training the student model.
arXiv Detail & Related papers (2023-03-27T09:16:10Z) - Self-Supervised Learning as a Means To Reduce the Need for Labeled Data
in Medical Image Analysis [64.4093648042484]
We use a dataset of chest X-ray images with bounding box labels for 13 different classes of anomalies.
We show that it is possible to achieve similar performance to a fully supervised model in terms of mean average precision and accuracy with only 60% of the labeled data.
arXiv Detail & Related papers (2022-06-01T09:20:30Z) - Robust Medical Image Classification from Noisy Labeled Data with Global
and Local Representation Guided Co-training [73.60883490436956]
We propose a novel collaborative training paradigm with global and local representation learning for robust medical image classification.
We employ the self-ensemble model with a noisy label filter to efficiently select the clean and noisy samples.
We also design a novel global and local representation learning scheme to implicitly regularize the networks to utilize noisy samples.
arXiv Detail & Related papers (2022-05-10T07:50:08Z) - LifeLonger: A Benchmark for Continual Disease Classification [59.13735398630546]
We introduce LifeLonger, a benchmark for continual disease classification on the MedMNIST collection.
Task and class incremental learning of diseases address the issue of classifying new samples without re-training the models from scratch.
Cross-domain incremental learning addresses the issue of dealing with datasets originating from different institutions while retaining the previously obtained knowledge.
arXiv Detail & Related papers (2022-04-12T12:25:05Z) - Learning from Multiple Expert Annotators for Enhancing Anomaly Detection
in Medical Image Analysis [0.31317409221921133]
In medical imaging analysis, multiple expert annotators often produce subjective estimates about "ground truth labels"
We propose a simple yet effective approach to combine annotations from multiple radiology experts for training a deep learning-based detector.
arXiv Detail & Related papers (2022-03-20T17:57:26Z) - About Explicit Variance Minimization: Training Neural Networks for
Medical Imaging With Limited Data Annotations [2.3204178451683264]
Variance Aware Training (VAT) method exploits this property by introducing the variance error into the model loss function.
We validate VAT on three medical imaging datasets from diverse domains and various learning objectives.
arXiv Detail & Related papers (2021-05-28T21:34:04Z) - Select-ProtoNet: Learning to Select for Few-Shot Disease Subtype
Prediction [55.94378672172967]
We focus on few-shot disease subtype prediction problem, identifying subgroups of similar patients.
We introduce meta learning techniques to develop a new model, which can extract the common experience or knowledge from interrelated clinical tasks.
Our new model is built upon a carefully designed meta-learner, called Prototypical Network, that is a simple yet effective meta learning machine for few-shot image classification.
arXiv Detail & Related papers (2020-09-02T02:50:30Z) - Confident Coreset for Active Learning in Medical Image Analysis [57.436224561482966]
We propose a novel active learning method, confident coreset, which considers both uncertainty and distribution for effectively selecting informative samples.
By comparative experiments on two medical image analysis tasks, we show that our method outperforms other active learning methods.
arXiv Detail & Related papers (2020-04-05T13:46:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.