Vision Transformers for femur fracture classification
- URL: http://arxiv.org/abs/2108.03414v1
- Date: Sat, 7 Aug 2021 10:12:42 GMT
- Title: Vision Transformers for femur fracture classification
- Authors: Leonardo Tanzi and Andrea Audisio and Giansalvo Cirrincione and
Alessandro Aprato and Enrico Vezzetti
- Abstract summary: The Vision Transformer (ViT) was able to correctly predict 83% of the test images.
Good results were obtained in sub-fractures with the largest and richest dataset ever.
- Score: 59.99241204074268
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Objectives: In recent years, the scientific community has focused on the
development of Computer-Aided Diagnosis (CAD) tools that could improve bone
fractures' classification. However, the results of the classification of
fractures in subtypes with the proposed datasets were far from optimal. This
paper proposes a very recent and outperforming deep learning technique, the
Vision Transformer (ViT), in order to improve the fracture classification, by
exploiting its self-attention mechanism.
Methods: 4207 manually annotated images were used and distributed, by
following the AO/OTA classification, in different fracture types, the largest
labeled dataset of proximal femur fractures used in literature. The ViT
architecture was used and compared with a classic Convolutional Neural Network
(CNN) and a multistage architecture composed by successive CNNs in cascade. To
demonstrate the reliability of this approach, 1) the attention maps were used
to visualize the most relevant areas of the images, 2) the performance of a
generic CNN and ViT was also compared through unsupervised learning techniques,
and 3) 11 specialists were asked to evaluate and classify 150 proximal femur
fractures' images with and without the help of the ViT.
Results: The ViT was able to correctly predict 83% of the test images.
Precision, recall and F1-score were 0.77 (CI 0.64-0.90), 0.76 (CI 0.62-0.91)
and 0.77 (CI 0.64-0.89), respectively. The average specialists' diagnostic
improvement was 29%.
Conclusions: This paper showed the potential of Transformers in bone fracture
classification. For the first time, good results were obtained in sub-fractures
with the largest and richest dataset ever.
Related papers
- Vision Transformer for Efficient Chest X-ray and Gastrointestinal Image
Classification [2.3293678240472517]
This study uses different CNNs and transformer-based methods with a wide range of data augmentation techniques.
We evaluated their performance on three medical image datasets from different modalities.
arXiv Detail & Related papers (2023-04-23T04:07:03Z) - Attention-based Saliency Maps Improve Interpretability of Pneumothorax
Classification [52.77024349608834]
To investigate chest radiograph (CXR) classification performance of vision transformers (ViT) and interpretability of attention-based saliency.
ViTs were fine-tuned for lung disease classification using four public data sets: CheXpert, Chest X-Ray 14, MIMIC CXR, and VinBigData.
ViTs had comparable CXR classification AUCs compared with state-of-the-art CNNs.
arXiv Detail & Related papers (2023-03-03T12:05:41Z) - Data-Efficient Vision Transformers for Multi-Label Disease
Classification on Chest Radiographs [55.78588835407174]
Vision Transformers (ViTs) have not been applied to this task despite their high classification performance on generic images.
ViTs do not rely on convolutions but on patch-based self-attention and in contrast to CNNs, no prior knowledge of local connectivity is present.
Our results show that while the performance between ViTs and CNNs is on par with a small benefit for ViTs, DeiTs outperform the former if a reasonably large data set is available for training.
arXiv Detail & Related papers (2022-08-17T09:07:45Z) - Predicting skull fractures via CNN with classification algorithms [0.0]
ResNet50 was developed to classify skull fractures from brain CT scans into three fracture categories.
It had the best overall F1-score of 96%, Hamming Score of 95%, Balanced accuracy Score of 94% & ROC AUC curve of 96% for the classification of skull fractures.
arXiv Detail & Related papers (2022-08-14T01:37:23Z) - Automated SSIM Regression for Detection and Quantification of Motion
Artefacts in Brain MR Images [54.739076152240024]
Motion artefacts in magnetic resonance brain images are a crucial issue.
The assessment of MR image quality is fundamental before proceeding with the clinical diagnosis.
An automated image quality assessment based on the structural similarity index (SSIM) regression has been proposed here.
arXiv Detail & Related papers (2022-06-14T10:16:54Z) - Osteoporosis Prescreening using Panoramic Radiographs through a Deep
Convolutional Neural Network with Attention Mechanism [65.70943212672023]
Deep convolutional neural network (CNN) with an attention module can detect osteoporosis on panoramic radiographs.
dataset of 70 panoramic radiographs (PRs) from 70 different subjects of age between 49 to 60 was used.
arXiv Detail & Related papers (2021-10-19T00:03:57Z) - Classification of Fracture and Normal Shoulder Bone X-Ray Images Using
Ensemble and Transfer Learning With Deep Learning Models Based on
Convolutional Neural Networks [0.0]
Various reasons cause shoulder fractures to occur, an area with wider and more varied range of movement than other joints in body.
Images in digital imaging and communications in medicine (DICOM) format are generated for shoulder via Xradiation (Xray), magnetic resonance imaging (MRI) or computed tomography (CT) devices.
Shoulder bone Xray images were classified and compared via deep learning models based on convolutional neural network (CNN) using transfer learning and ensemble learning.
arXiv Detail & Related papers (2021-01-31T19:20:04Z) - A Deep Learning Study on Osteosarcoma Detection from Histological Images [6.341765152919201]
The most common type of primary malignant bone tumor is osteosarcoma.
CNNs can significantly decrease surgeon's workload and make a better prognosis of patient conditions.
CNNs need to be trained on a large amount of data in order to achieve a more trustworthy performance.
arXiv Detail & Related papers (2020-11-02T18:16:17Z) - Classification of COVID-19 in CT Scans using Multi-Source Transfer
Learning [91.3755431537592]
We propose the use of Multi-Source Transfer Learning to improve upon traditional Transfer Learning for the classification of COVID-19 from CT scans.
With our multi-source fine-tuning approach, our models outperformed baseline models fine-tuned with ImageNet.
Our best performing model was able to achieve an accuracy of 0.893 and a Recall score of 0.897, outperforming its baseline Recall score by 9.3%.
arXiv Detail & Related papers (2020-09-22T11:53:06Z) - Joint Learning of Vessel Segmentation and Artery/Vein Classification
with Post-processing [27.825969553813092]
Vessel segmentation and artery/vein classification provide various information on potential disorders.
We adopt a UNet-based model, SeqNet, to accurately segment vessels from the background and make prediction on the vessel type.
Our experiments show that our method improves AUC to 0.98 for segmentation and the accuracy to 0.92 in classification over DRIVE dataset.
arXiv Detail & Related papers (2020-05-27T13:06:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.