Automated Labeling of German Chest X-Ray Radiology Reports using Deep
Learning
- URL: http://arxiv.org/abs/2306.05997v2
- Date: Fri, 7 Jul 2023 11:36:11 GMT
- Title: Automated Labeling of German Chest X-Ray Radiology Reports using Deep
Learning
- Authors: Alessandro Wollek, Philip Haitzer, Thomas Sedlmeyr, Sardi Hyska,
Johannes Rueckel, Bastian Sabel, Michael Ingrisch, Tobias Lasser
- Abstract summary: We propose a deep learning-based CheXpert label prediction model, pre-trained on reports labeled by a rule-based German CheXpert model.
Our results demonstrate the effectiveness of our approach, which significantly outperformed the rule-based model on all three tasks.
- Score: 50.591267188664666
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Radiologists are in short supply globally, and deep learning models offer a
promising solution to address this shortage as part of clinical
decision-support systems. However, training such models often requires
expensive and time-consuming manual labeling of large datasets. Automatic label
extraction from radiology reports can reduce the time required to obtain
labeled datasets, but this task is challenging due to semantically similar
words and missing annotated data. In this work, we explore the potential of
weak supervision of a deep learning-based label prediction model, using a
rule-based labeler. We propose a deep learning-based CheXpert label prediction
model, pre-trained on reports labeled by a rule-based German CheXpert model and
fine-tuned on a small dataset of manually labeled reports. Our results
demonstrate the effectiveness of our approach, which significantly outperformed
the rule-based model on all three tasks. Our findings highlight the benefits of
employing deep learning-based models even in scenarios with sparse data and the
use of the rule-based labeler as a tool for weak supervision.
Related papers
- Few Shot Rationale Generation using Self-Training with Dual Teachers [4.91890875296663]
Self-rationalizing models that also generate a free-text explanation for their predicted labels are an important tool to build trustworthy AI applications.
We introduce a novel dual-teacher learning framework, which learns two specialized teacher models for task prediction and rationalization.
We formulate a new loss function, Masked Label Regularization (MLR) which promotes explanations to be strongly conditioned on predicted labels.
arXiv Detail & Related papers (2023-06-05T23:57:52Z) - German CheXpert Chest X-ray Radiology Report Labeler [50.591267188664666]
This study aimed to develop an algorithm to automatically extract annotations for chest X-ray classification models from German thoracic radiology reports.
Results showed that automated label extraction can reduce time spent on manual labeling and improve overall modeling performance.
arXiv Detail & Related papers (2023-06-05T11:01:58Z) - A Benchmark Generative Probabilistic Model for Weak Supervised Learning [2.0257616108612373]
Weak Supervised Learning approaches have been developed to alleviate the annotation burden.
We show that latent variable models (PLVMs) achieve state-of-the-art performance across four datasets.
arXiv Detail & Related papers (2023-03-31T07:06:24Z) - Clinically Acceptable Segmentation of Organs at Risk in Cervical Cancer
Radiation Treatment from Clinically Available Annotations [0.0]
We present an approach to learn a deep learning model for the automatic segmentation of Organs at Risk (OARs) in cervical cancer radiation treatment.
We employ simples for automatic data cleaning to minimize data inhomogeneity, label noise, and missing annotations.
We develop a semi-supervised learning approach utilizing a teacher-student setup, annotation imputation, and uncertainty-guided training to learn in presence of missing annotations.
arXiv Detail & Related papers (2023-02-21T13:24:40Z) - Learned Label Aggregation for Weak Supervision [8.819582879892762]
We propose a data programming approach that aggregates weak supervision signals to generate labeled data easily.
The quality of the generated labels depends on a label aggregation model that aggregates all noisy labels from all LFs to infer the ground-truth labels.
We show the model can be trained using synthetically generated data and design an effective architecture for the model.
arXiv Detail & Related papers (2022-07-27T14:36:35Z) - Self-Supervised Learning as a Means To Reduce the Need for Labeled Data
in Medical Image Analysis [64.4093648042484]
We use a dataset of chest X-ray images with bounding box labels for 13 different classes of anomalies.
We show that it is possible to achieve similar performance to a fully supervised model in terms of mean average precision and accuracy with only 60% of the labeled data.
arXiv Detail & Related papers (2022-06-01T09:20:30Z) - Learning from Noisy Labels for Entity-Centric Information Extraction [17.50856935207308]
We propose a simple co-regularization framework for entity-centric information extraction.
These models are jointly optimized with task-specific loss, and are regularized to generate similar predictions.
In the end, we can take any of the trained models for inference.
arXiv Detail & Related papers (2021-04-17T22:49:12Z) - Uncertainty-aware Self-training for Text Classification with Few Labels [54.13279574908808]
We study self-training as one of the earliest semi-supervised learning approaches to reduce the annotation bottleneck.
We propose an approach to improve self-training by incorporating uncertainty estimates of the underlying neural network.
We show our methods leveraging only 20-30 labeled samples per class for each task for training and for validation can perform within 3% of fully supervised pre-trained language models.
arXiv Detail & Related papers (2020-06-27T08:13:58Z) - Self-Training with Improved Regularization for Sample-Efficient Chest
X-Ray Classification [80.00316465793702]
We present a deep learning framework that enables robust modeling in challenging scenarios.
Our results show that using 85% lesser labeled data, we can build predictive models that match the performance of classifiers trained in a large-scale data setting.
arXiv Detail & Related papers (2020-05-03T02:36:00Z) - Deep k-NN for Noisy Labels [55.97221021252733]
We show that a simple $k$-nearest neighbor-based filtering approach on the logit layer of a preliminary model can remove mislabeled data and produce more accurate models than many recently proposed methods.
arXiv Detail & Related papers (2020-04-26T05:15:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.