DICOM Imaging Router: An Open Deep Learning Framework for Classification
of Body Parts from DICOM X-ray Scans
- URL: http://arxiv.org/abs/2108.06490v2
- Date: Tue, 17 Aug 2021 04:01:20 GMT
- Title: DICOM Imaging Router: An Open Deep Learning Framework for Classification
of Body Parts from DICOM X-ray Scans
- Authors: Hieu H. Pham, Dung V. Do, Ha Q. Nguyen
- Abstract summary: We introduce a DICOM Imaging Router that deploys deep CNNs for categorizing unknown DICOM X-ray images into five anatomical groups.
We trained a set of state-of-the-art deep CNNs using a training set of 11,263 images and showed superior performance in classifying the body parts.
- Score: 0.618778092044887
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: X-ray imaging in DICOM format is the most commonly used imaging modality in
clinical practice, resulting in vast, non-normalized databases. This leads to
an obstacle in deploying AI solutions for analyzing medical images, which often
requires identifying the right body part before feeding the image into a
specified AI model. This challenge raises the need for an automated and
efficient approach to classifying body parts from X-ray scans. Unfortunately,
to the best of our knowledge, there is no open tool or framework for this task
to date. To fill this lack, we introduce a DICOM Imaging Router that deploys
deep CNNs for categorizing unknown DICOM X-ray images into five anatomical
groups: abdominal, adult chest, pediatric chest, spine, and others. To this
end, a large-scale X-ray dataset consisting of 16,093 images has been collected
and manually classified. We then trained a set of state-of-the-art deep CNNs
using a training set of 11,263 images. These networks were then evaluated on an
independent test set of 2,419 images and showed superior performance in
classifying the body parts. Specifically, our best performing model achieved a
recall of 0.982 (95% CI, 0.977-0.988), a precision of 0.985 (95% CI,
0.975-0.989) and a F1-score of 0.981 (95% CI, 0.976-0.987), whilst requiring
less computation for inference (0.0295 second per image). Our external validity
on 1,000 X-ray images shows the robustness of the proposed approach across
hospitals. These remarkable performances indicate that deep CNNs can accurately
and effectively differentiate human body parts from X-ray scans, thereby
providing potential benefits for a wide range of applications in clinical
settings. The dataset, codes, and trained deep learning models from this study
will be made publicly available on our project website at https://vindr.ai/.
Related papers
- TotalSegmentator MRI: Sequence-Independent Segmentation of 59 Anatomical Structures in MR images [62.53931644063323]
In this study we extended the capabilities of TotalSegmentator to MR images.
We trained an nnU-Net segmentation algorithm on this dataset and calculated similarity coefficients (Dice) to evaluate the model's performance.
The model significantly outperformed two other publicly available segmentation models (Dice score 0.824 versus 0.762; p0.001 and 0.762 versus 0.542; p)
arXiv Detail & Related papers (2024-05-29T20:15:54Z) - Medical Image Analysis for Detection, Treatment and Planning of Disease using Artificial Intelligence Approaches [1.6505331001136514]
A framework for the segmentation of X-ray images using artificial intelligence techniques has been discussed.
The proposed approach performs better in all respect of well-known parameters with 16 batch size and 50 epochs.
The value of validation accuracy, precision, and recall of SegNet and Residual Unet models are 0.9815, 0.9699, 0.9574, and 0.9901, 0.9864, 0.9750 respectively.
arXiv Detail & Related papers (2024-05-18T13:43:43Z) - CT-GLIP: 3D Grounded Language-Image Pretraining with CT Scans and Radiology Reports for Full-Body Scenarios [53.94122089629544]
We introduce CT-GLIP (Grounded Language-Image Pretraining with CT scans), a novel method that constructs organ-level image-text pairs to enhance multimodal contrastive learning.
Our method, trained on a multimodal CT dataset comprising 44,011 organ-level vision-text pairs from 17,702 patients across 104 organs, demonstrates it can identify organs and abnormalities in a zero-shot manner using natural languages.
arXiv Detail & Related papers (2024-04-23T17:59:01Z) - Data-Efficient Vision Transformers for Multi-Label Disease
Classification on Chest Radiographs [55.78588835407174]
Vision Transformers (ViTs) have not been applied to this task despite their high classification performance on generic images.
ViTs do not rely on convolutions but on patch-based self-attention and in contrast to CNNs, no prior knowledge of local connectivity is present.
Our results show that while the performance between ViTs and CNNs is on par with a small benefit for ViTs, DeiTs outperform the former if a reasonably large data set is available for training.
arXiv Detail & Related papers (2022-08-17T09:07:45Z) - COVID-19 Severity Classification on Chest X-ray Images [0.0]
In this work, we classify covid images based on the severity of the infection.
The ResNet-50 model produced remarkable classification results in terms of accuracy 95%, recall (0.94), and F1-Score (0.92), and precision (0.91)
arXiv Detail & Related papers (2022-05-25T12:01:03Z) - Preservation of High Frequency Content for Deep Learning-Based Medical
Image Classification [74.84221280249876]
An efficient analysis of large amounts of chest radiographs can aid physicians and radiologists.
We propose a novel Discrete Wavelet Transform (DWT)-based method for the efficient identification and encoding of visual information.
arXiv Detail & Related papers (2022-05-08T15:29:54Z) - Towards Clinical Practice: Design and Implementation of Convolutional
Neural Network-Based Assistive Diagnosis System for COVID-19 Case Detection
from Chest X-Ray Images [0.0]
This study presents a real-world implementation of a convolutional neural network (CNN) based Carebot Covid app to detect COVID-19 from chest X-ray (CXR) images.
The results of this study show that the deep learning model based on DenseNet and ResNet architecture can detect SARS-CoV-2 from CXR images with precision of 0.981, recall of 0.962 and AP of 0.993.
arXiv Detail & Related papers (2022-03-20T16:44:20Z) - Vision Transformers for femur fracture classification [59.99241204074268]
The Vision Transformer (ViT) was able to correctly predict 83% of the test images.
Good results were obtained in sub-fractures with the largest and richest dataset ever.
arXiv Detail & Related papers (2021-08-07T10:12:42Z) - Image Embedding and Model Ensembling for Automated Chest X-Ray
Interpretation [0.0]
We present and study several machine learning approaches to develop automated Chest X-ray diagnostic models.
In particular, we trained several Convolutional Neural Networks (CNN) on the CheXpert dataset.
We used the trained CNNs to compute embeddings of the CXR images, in order to train two sets of tree-based classifiers from them.
arXiv Detail & Related papers (2021-05-05T14:48:59Z) - Fused Deep Convolutional Neural Network for Precision Diagnosis of
COVID-19 Using Chest X-Ray Images [0.0]
We propose a computer-aided diagnosis (CAD) to accurately classify chest X-ray scans of COVID-19 and normal subjects by fine-tuning several neural networks.
Using k-fold cross-validation and a bagging ensemble, we achieve an accuracy of 99.7% and a sensitivity of 100%.
arXiv Detail & Related papers (2020-09-15T02:27:20Z) - Y-Net for Chest X-Ray Preprocessing: Simultaneous Classification of
Geometry and Segmentation of Annotations [70.0118756144807]
This work introduces a general pre-processing step for chest x-ray input into machine learning algorithms.
A modified Y-Net architecture based on the VGG11 encoder is used to simultaneously learn geometric orientation and segmentation of radiographs.
Results were evaluated by expert clinicians, with acceptable geometry in 95.8% and annotation mask in 96.2%, compared to 27.0% and 34.9% respectively in control images.
arXiv Detail & Related papers (2020-05-08T02:16:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.