Shifted Windows Transformers for Medical Image Quality Assessment
- URL: http://arxiv.org/abs/2208.06034v1
- Date: Thu, 11 Aug 2022 20:39:26 GMT
- Title: Shifted Windows Transformers for Medical Image Quality Assessment
- Authors: Caner Ozer, Arda Guler, Aysel Turkvatan Cansever, Deniz Alis, Ercan
Karaarslan, Ilkay Oksuz
- Abstract summary: CNN-based approaches are used to assess the image quality, but their performance can still be improved in terms of accuracy.
In this work, we approach this problem by using Swin Transformer, which improves the poor-quality image classification performance.
To the best of our knowledge, our study is the first vision transformer application for medical image quality assessment.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: To maintain a standard in a medical imaging study, images should have
necessary image quality for potential diagnostic use. Although CNN-based
approaches are used to assess the image quality, their performance can still be
improved in terms of accuracy. In this work, we approach this problem by using
Swin Transformer, which improves the poor-quality image classification
performance that causes the degradation in medical image quality. We test our
approach on Foreign Object Classification problem on Chest X-Rays (Object-CXR)
and Left Ventricular Outflow Tract Classification problem on Cardiac MRI with a
four-chamber view (LVOT). While we obtain a classification accuracy of 87.1%
and 95.48% on the Object-CXR and LVOT datasets, our experimental results
suggest that the use of Swin Transformer improves the Object-CXR classification
performance while obtaining a comparable performance for the LVOT dataset. To
the best of our knowledge, our study is the first vision transformer
application for medical image quality assessment.
Related papers
- DCT-HistoTransformer: Efficient Lightweight Vision Transformer with DCT Integration for histopathological image analysis [0.0]
We introduce a novel lightweight breast cancer classification approach using Vision Transformers (ViTs)
By incorporating parallel processing pathways for Discrete Cosine Transform (DCT) Attention and MobileConv, we convert image data from the spatial domain to the frequency domain to utilize the benefits such as filtering out high frequencies in the image.
Our proposed model achieves an accuracy of 96.00% $pm$ 0.48% for binary classification and 87.85% $pm$ 0.93% for multiclass classification, which is comparable to state-of-the-art models.
arXiv Detail & Related papers (2024-10-24T21:16:56Z) - Blind CT Image Quality Assessment Using DDPM-derived Content and
Transformer-based Evaluator [9.626301541775998]
Blind image quality assessment (BIQA) strives to evaluate perceptual quality in alignment with what radiologists perceive.
In this study, we introduce an innovative BIQA metric that emulates the active inference process of IGM.
We won the second place in the MICCAI 2023 low-dose computed tomography perceptual image quality assessment grand challenge.
arXiv Detail & Related papers (2023-10-04T19:13:16Z) - Performance of GAN-based augmentation for deep learning COVID-19 image
classification [57.1795052451257]
The biggest challenge in the application of deep learning to the medical domain is the availability of training data.
Data augmentation is a typical methodology used in machine learning when confronted with a limited data set.
In this work, a StyleGAN2-ADA model of Generative Adversarial Networks is trained on the limited COVID-19 chest X-ray image set.
arXiv Detail & Related papers (2023-04-18T15:39:58Z) - Explainable Image Quality Assessment for Medical Imaging [0.0]
Poor-quality medical images may lead to misdiagnosis.
We propose an explainable image quality assessment system and validate our idea on two different objectives.
We apply a variety of techniques to measure the faithfulness of the saliency detectors.
We show that NormGrad has significant gains over other saliency detectors by reaching a repeated Pointing Game score of 0.853 for Object-CXR and 0.611 for LVOT datasets.
arXiv Detail & Related papers (2023-03-25T14:18:39Z) - Optimising Chest X-Rays for Image Analysis by Identifying and Removing
Confounding Factors [49.005337470305584]
During the COVID-19 pandemic, the sheer volume of imaging performed in an emergency setting for COVID-19 diagnosis has resulted in a wide variability of clinical CXR acquisitions.
The variable quality of clinically-acquired CXRs within publicly available datasets could have a profound effect on algorithm performance.
We propose a simple and effective step-wise approach to pre-processing a COVID-19 chest X-ray dataset to remove undesired biases.
arXiv Detail & Related papers (2022-08-22T13:57:04Z) - Automated SSIM Regression for Detection and Quantification of Motion
Artefacts in Brain MR Images [54.739076152240024]
Motion artefacts in magnetic resonance brain images are a crucial issue.
The assessment of MR image quality is fundamental before proceeding with the clinical diagnosis.
An automated image quality assessment based on the structural similarity index (SSIM) regression has been proposed here.
arXiv Detail & Related papers (2022-06-14T10:16:54Z) - Preservation of High Frequency Content for Deep Learning-Based Medical
Image Classification [74.84221280249876]
An efficient analysis of large amounts of chest radiographs can aid physicians and radiologists.
We propose a novel Discrete Wavelet Transform (DWT)-based method for the efficient identification and encoding of visual information.
arXiv Detail & Related papers (2022-05-08T15:29:54Z) - RFormer: Transformer-based Generative Adversarial Network for Real
Fundus Image Restoration on A New Clinical Benchmark [8.109057397954537]
Ophthalmologists have used fundus images to screen and diagnose eye diseases.
Low-quality (LQ) degraded fundus images easily lead to uncertainty in clinical screening and generally increase the risk of misdiagnosis.
We propose a novel Transformer-based Generative Adversarial Network (RFormer) to restore the real degradation of clinical fundus images.
arXiv Detail & Related papers (2022-01-03T03:56:58Z) - Learning Conditional Knowledge Distillation for Degraded-Reference Image
Quality Assessment [157.1292674649519]
We propose a practical solution named degraded-reference IQA (DR-IQA)
DR-IQA exploits the inputs of IR models, degraded images, as references.
Our results can even be close to the performance of full-reference settings.
arXiv Detail & Related papers (2021-08-18T02:35:08Z) - Residual Networks based Distortion Classification and Ranking for
Laparoscopic Image Quality Assessment [12.374294852377382]
Laparoscopic images and videos are often affected by different types of distortion like noise, smoke, blur and nonuniform illumination.
We propose in this paper to formulate the image quality assessment task as a multi-label classification problem taking into account both the type as well as the severity level (or rank) of distortions.
The obtained results on a laparoscopic image dataset show the efficiency of the proposed approach.
arXiv Detail & Related papers (2021-06-12T14:26:11Z) - Classification of COVID-19 in CT Scans using Multi-Source Transfer
Learning [91.3755431537592]
We propose the use of Multi-Source Transfer Learning to improve upon traditional Transfer Learning for the classification of COVID-19 from CT scans.
With our multi-source fine-tuning approach, our models outperformed baseline models fine-tuned with ImageNet.
Our best performing model was able to achieve an accuracy of 0.893 and a Recall score of 0.897, outperforming its baseline Recall score by 9.3%.
arXiv Detail & Related papers (2020-09-22T11:53:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.