Vision Transformer using Low-level Chest X-ray Feature Corpus for
COVID-19 Diagnosis and Severity Quantification
- URL: http://arxiv.org/abs/2104.07235v1
- Date: Thu, 15 Apr 2021 04:54:48 GMT
- Title: Vision Transformer using Low-level Chest X-ray Feature Corpus for
COVID-19 Diagnosis and Severity Quantification
- Authors: Sangjoon Park, Gwanghyun Kim, Yujin Oh, Joon Beom Seo, Sang Min Lee,
Jin Hwan Kim, Sungjun Moon, Jae-Kwang Lim, Jong Chul Ye
- Abstract summary: We propose a novel Vision Transformer that utilizes low-level CXR feature corpus obtained from a backbone network.
The backbone network is first trained with large public datasets to detect common abnormal findings.
Then, the embedded features from the backbone network are used as corpora for a Transformer model for the diagnosis and the severity quantification of COVID-19.
- Score: 25.144248675578286
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Developing a robust algorithm to diagnose and quantify the severity of
COVID-19 using Chest X-ray (CXR) requires a large number of well-curated
COVID-19 datasets, which is difficult to collect under the global COVID-19
pandemic. On the other hand, CXR data with other findings are abundant. This
situation is ideally suited for the Vision Transformer (ViT) architecture,
where a lot of unlabeled data can be used through structural modeling by the
self-attention mechanism. However, the use of existing ViT is not optimal,
since feature embedding through direct patch flattening or ResNet backbone in
the standard ViT is not intended for CXR. To address this problem, here we
propose a novel Vision Transformer that utilizes low-level CXR feature corpus
obtained from a backbone network that extracts common CXR findings.
Specifically, the backbone network is first trained with large public datasets
to detect common abnormal findings such as consolidation, opacity, edema, etc.
Then, the embedded features from the backbone network are used as corpora for a
Transformer model for the diagnosis and the severity quantification of
COVID-19. We evaluate our model on various external test datasets from totally
different institutions to evaluate the generalization capability. The
experimental results confirm that our model can achieve the state-of-the-art
performance in both diagnosis and severity quantification tasks with superior
generalization capability, which are sine qua non of widespread deployment.
Related papers
- Towards a Benchmark for Colorectal Cancer Segmentation in Endorectal Ultrasound Videos: Dataset and Model Development [59.74920439478643]
In this paper, we collect and annotated the first benchmark dataset that covers diverse ERUS scenarios.
Our ERUS-10K dataset comprises 77 videos and 10,000 high-resolution annotated frames.
We introduce a benchmark model for colorectal cancer segmentation, named the Adaptive Sparse-context TRansformer (ASTR)
arXiv Detail & Related papers (2024-08-19T15:04:42Z) - Multi-Scale Feature Fusion using Parallel-Attention Block for COVID-19
Chest X-ray Diagnosis [2.15242029196761]
Under the global COVID-19 crisis, accurate diagnosis of COVID-19 from Chest X-ray (CXR) images is critical.
We propose a novel multi-feature fusion network using parallel attention blocks to fuse the original CXR images and local-phase feature-enhanced CXR images at multi-scales.
arXiv Detail & Related papers (2023-04-25T16:56:12Z) - Performance of GAN-based augmentation for deep learning COVID-19 image
classification [57.1795052451257]
The biggest challenge in the application of deep learning to the medical domain is the availability of training data.
Data augmentation is a typical methodology used in machine learning when confronted with a limited data set.
In this work, a StyleGAN2-ADA model of Generative Adversarial Networks is trained on the limited COVID-19 chest X-ray image set.
arXiv Detail & Related papers (2023-04-18T15:39:58Z) - Vision Transformer-based Model for Severity Quantification of Lung
Pneumonia Using Chest X-ray Images [11.12596879975844]
We present a Vision Transformer-based neural network model that relies on a small number of trainable parameters to quantify the severity of COVID-19 and other lung diseases.
Our model can provide peak performance in quantifying severity with high generalizability at a relatively low computational cost.
arXiv Detail & Related papers (2023-03-18T12:38:23Z) - Optimising Chest X-Rays for Image Analysis by Identifying and Removing
Confounding Factors [49.005337470305584]
During the COVID-19 pandemic, the sheer volume of imaging performed in an emergency setting for COVID-19 diagnosis has resulted in a wide variability of clinical CXR acquisitions.
The variable quality of clinically-acquired CXRs within publicly available datasets could have a profound effect on algorithm performance.
We propose a simple and effective step-wise approach to pre-processing a COVID-19 chest X-ray dataset to remove undesired biases.
arXiv Detail & Related papers (2022-08-22T13:57:04Z) - The pitfalls of using open data to develop deep learning solutions for
COVID-19 detection in chest X-rays [64.02097860085202]
Deep learning models have been developed to identify COVID-19 from chest X-rays.
Results have been exceptional when training and testing on open-source data.
Data analysis and model evaluations show that the popular open-source dataset COVIDx is not representative of the real clinical problem.
arXiv Detail & Related papers (2021-09-14T10:59:11Z) - Severity Quantification and Lesion Localization of COVID-19 on CXR using
Vision Transformer [25.144248675578286]
Under the global pandemic of COVID-19, building an automated framework that quantifies the severity of COVID-19 has become increasingly important.
We propose a novel Vision Transformer tailored for both quantification of the severity and clinically applicable localization of the COVID-19 related lesions.
Our model is trained in a weakly-supervised manner to generate the full probability maps from weak array-based labels.
arXiv Detail & Related papers (2021-03-12T03:17:19Z) - Vision Transformer for COVID-19 CXR Diagnosis using Chest X-ray Feature
Corpus [25.144248675578286]
Under the global COVID-19 crisis, developing robust diagnosis algorithm for COVID-19 using CXR is hampered by the lack of well-curated COVID-19 data set.
We propose a novel vision Transformer by using the low-level CXR feature corpus that are obtained to extract the abnormal CXR features.
arXiv Detail & Related papers (2021-03-12T03:07:27Z) - Many-to-One Distribution Learning and K-Nearest Neighbor Smoothing for
Thoracic Disease Identification [83.6017225363714]
deep learning has become the most powerful computer-aided diagnosis technology for improving disease identification performance.
For chest X-ray imaging, annotating large-scale data requires professional domain knowledge and is time-consuming.
In this paper, we propose many-to-one distribution learning (MODL) and K-nearest neighbor smoothing (KNNS) methods to improve a single model's disease identification performance.
arXiv Detail & Related papers (2021-02-26T02:29:30Z) - End-2-End COVID-19 Detection from Breath & Cough Audio [68.41471917650571]
We demonstrate the first attempt to diagnose COVID-19 using end-to-end deep learning from a crowd-sourced dataset of audio samples.
We introduce a novel modelling strategy using a custom deep neural network to diagnose COVID-19 from a joint breath and cough representation.
arXiv Detail & Related papers (2021-01-07T01:13:00Z) - Intra-model Variability in COVID-19 Classification Using Chest X-ray
Images [0.0]
We quantify baseline performance metrics and variability for COVID-19 detection in chest x-ray for 12 common deep learning architectures.
Best performing models achieve a false negative rate of 3 out of 20 for detecting COVID-19 in a hold-out set.
arXiv Detail & Related papers (2020-04-30T21:20:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.