Developing A Visual-Interactive Interface for Electronic Health Record
Labeling: An Explainable Machine Learning Approach
- URL: http://arxiv.org/abs/2209.12778v2
- Date: Fri, 2 Jun 2023 05:23:35 GMT
- Title: Developing A Visual-Interactive Interface for Electronic Health Record
Labeling: An Explainable Machine Learning Approach
- Authors: Donlapark Ponnoprat, Parichart Pattarapanitchai, Phimphaka Taninpong,
Suthep Suantai, Natthanaphop Isaradech, Thiraphat Tanphiriyakun
- Abstract summary: We introduce Explainable Labeling Assistant (XLabel) a new visual-interactive tool for data labeling.
XLabel uses Explainable Boosting Machine (EBM) to classify the labels of each data point and visualizes heatmaps of EBM's explanations.
Our experiments show that 1) XLabel helps reduce the number of labeling actions, 2) EBM as an explainable classifier is as accurate as other well-known machine learning models, and 3) even when more than 40% of the records were intentionally mislabeled, EBM could recall the correct labels of more than 90% of these records.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Labeling a large number of electronic health records is expensive and time
consuming, and having a labeling assistant tool can significantly reduce
medical experts' workload. Nevertheless, to gain the experts' trust, the tool
must be able to explain the reasons behind its outputs. Motivated by this, we
introduce Explainable Labeling Assistant (XLabel) a new visual-interactive tool
for data labeling. At a high level, XLabel uses Explainable Boosting Machine
(EBM) to classify the labels of each data point and visualizes heatmaps of
EBM's explanations. As a case study, we use XLabel to help medical experts
label electronic health records with four common non-communicable diseases
(NCDs). Our experiments show that 1) XLabel helps reduce the number of labeling
actions, 2) EBM as an explainable classifier is as accurate as other well-known
machine learning models outperforms a rule-based model used by NCD experts, and
3) even when more than 40% of the records were intentionally mislabeled, EBM
could recall the correct labels of more than 90% of these records.
Related papers
- Automatic Labels are as Effective as Manual Labels in Biomedical Images Classification with Deep Learning [32.71326861917332]
One of the main limitations to train deep learning algorithms to perform a specific task is the need for medical experts to label data.
This paper aims to investigate under which circumstances automatic labels can be adopted to train a DL model on the classification of Whole Slide Images (WSI)
The application of the Semantic Knowledge Extractor Tool (SKET) algorithm to generate automatic labels leads to performance comparable to the one obtained with manual labels.
arXiv Detail & Related papers (2024-06-20T14:20:50Z) - Auto-Generating Weak Labels for Real & Synthetic Data to Improve Label-Scarce Medical Image Segmentation [5.368714143438489]
We present a new approach to overcome the hurdle of costly medical image labeling by leveraging foundation models like Segment Anything Model (SAM) and its medical alternate MedSAM.
Our pipeline has the ability to generate weak labels for any unlabeled medical image and subsequently use it to augment label-scarce datasets.
This automation eliminates the manual prompting step in MedSAM, creating a streamlined process for generating labels for both real and synthetic images.
arXiv Detail & Related papers (2024-04-25T20:47:08Z) - Automated Labeling of German Chest X-Ray Radiology Reports using Deep
Learning [50.591267188664666]
We propose a deep learning-based CheXpert label prediction model, pre-trained on reports labeled by a rule-based German CheXpert model.
Our results demonstrate the effectiveness of our approach, which significantly outperformed the rule-based model on all three tasks.
arXiv Detail & Related papers (2023-06-09T16:08:35Z) - Imprecise Label Learning: A Unified Framework for Learning with Various Imprecise Label Configurations [91.67511167969934]
imprecise label learning (ILL) is a framework for the unification of learning with various imprecise label configurations.
We demonstrate that ILL can seamlessly adapt to partial label learning, semi-supervised learning, noisy label learning, and, more importantly, a mixture of these settings.
arXiv Detail & Related papers (2023-05-22T04:50:28Z) - Self-Supervised Learning as a Means To Reduce the Need for Labeled Data
in Medical Image Analysis [64.4093648042484]
We use a dataset of chest X-ray images with bounding box labels for 13 different classes of anomalies.
We show that it is possible to achieve similar performance to a fully supervised model in terms of mean average precision and accuracy with only 60% of the labeled data.
arXiv Detail & Related papers (2022-06-01T09:20:30Z) - OneLabeler: A Flexible System for Building Data Labeling Tools [48.15772261649084]
OneLabeler supports configuration and composition of common software modules to build data labeling tools.
A user study with developers provides evidence that OneLabeler supports efficient building of diverse data labeling tools.
arXiv Detail & Related papers (2022-03-27T07:22:36Z) - Label-Assemble: Leveraging Multiple Datasets with Partial Labels [68.46767639240564]
"Label-Assemble" aims to unleash the full potential of partial labels from an assembly of public datasets.
We discovered that learning from negative examples facilitates both computer-aided disease diagnosis and detection.
arXiv Detail & Related papers (2021-09-25T02:48:17Z) - ECLARE: Extreme Classification with Label Graph Correlations [13.429436351837653]
This paper presents ECLARE, a scalable deep learning architecture that incorporates not only label text, but also label correlations, to offer accurate real-time predictions within a few milliseconds.
ECLARE offers predictions that are 2 to 14% more accurate on both publicly available benchmark datasets as well as proprietary datasets for a related products recommendation task sourced from the Bing search engine.
arXiv Detail & Related papers (2021-07-31T15:13:13Z) - A Study on the Autoregressive and non-Autoregressive Multi-label
Learning [77.11075863067131]
We propose a self-attention based variational encoder-model to extract the label-label and label-feature dependencies jointly.
Our model can therefore be used to predict all labels in parallel while still including both label-label and label-feature dependencies.
arXiv Detail & Related papers (2020-12-03T05:41:44Z) - Paying Per-label Attention for Multi-label Extraction from Radiology
Reports [1.9601378412924186]
We tackle the automated extraction of structured labels from head CT reports for imaging of suspected stroke patients.
We propose a set of 31 labels which correspond to radiographic findings and clinical impressions related to neurological abnormalities.
We are able to robustly extract many labels with a single model, classified according to the radiologist's reporting.
arXiv Detail & Related papers (2020-07-31T16:11:09Z) - CheXbert: Combining Automatic Labelers and Expert Annotations for
Accurate Radiology Report Labeling Using BERT [6.458158112222296]
We introduce a BERT-based approach to medical image report labeling.
We demonstrate superior performance of a biomedically pretrained BERT model first trained on annotations of a rule-based labeler.
We find that our final model, CheXbert, is able to outperform the previous best rules-based labeler with statistical significance.
arXiv Detail & Related papers (2020-04-20T09:46:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.