Artificially Generated Visual Scanpath Improves Multi-label Thoracic Disease Classification in Chest X-Ray Images
- URL: http://arxiv.org/abs/2503.00657v1
- Date: Sat, 01 Mar 2025 23:13:29 GMT
- Title: Artificially Generated Visual Scanpath Improves Multi-label Thoracic Disease Classification in Chest X-Ray Images
- Authors: Ashish Verma, Aupendu Kar, Krishnendu Ghosh, Sobhan Kanti Dhara, Debashis Sen, Prabir Kumar Biswas,
- Abstract summary: Expert radiologists visually scan Chest X-Ray (CXR) images, sequentially fixating on anatomical structures to perform disease diagnosis.<n>Recorded visual scanpaths of radiologists on CXR images can be used for the said purpose.<n>This paper proposes to generate effective artificial visual scanpaths using a visual scanpath prediction model for CXR images.
- Score: 17.040556955838138
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Expert radiologists visually scan Chest X-Ray (CXR) images, sequentially fixating on anatomical structures to perform disease diagnosis. An automatic multi-label classifier of diseases in CXR images can benefit by incorporating aspects of the radiologists' approach. Recorded visual scanpaths of radiologists on CXR images can be used for the said purpose. But, such scanpaths are not available for most CXR images, which creates a gap even for modern deep learning based classifiers. This paper proposes to mitigate this gap by generating effective artificial visual scanpaths using a visual scanpath prediction model for CXR images. Further, a multi-class multi-label classifier framework is proposed that uses a generated scanpath and visual image features to classify diseases in CXR images. While the scanpath predictor is based on a recurrent neural network, the multi-label classifier involves a novel iterative sequential model with an attention module. We show that our scanpath predictor generates human-like visual scanpaths. We also demonstrate that the use of artificial visual scanpaths improves multi-class multi-label disease classification results on CXR images. The above observations are made from experiments involving around 0.2 million CXR images from 2 widely-used datasets considering the multi-label classification of 14 pathological findings. Code link: https://github.com/ashishverma03/SDC
Related papers
- Visual Prompt Engineering for Vision Language Models in Radiology [0.17183214167143138]
Contrastive Language-Image Pretraining (CLIPP) offers a solution by enabling zero-shot classification through large-scale pretraining.<n>Visual markers improve AUROC2013$ by up to 0.185, highlighting their effectiveness enhancing classification performance.<n>We release our code and preprocessing pipeline, providing a reference point for future work on localized classification in medical imaging.
arXiv Detail & Related papers (2024-08-28T13:53:27Z) - LeDNet: Localization-enabled Deep Neural Network for Multi-Label Radiography Image Classification [0.1227734309612871]
Multi-label radiography image classification has long been a topic of interest in neural networks research.
We will use the chest x-ray images to detect thoracic diseases for this purpose.
We propose a combination of localization and deep learning algorithms called LeDNet to predict thoracic diseases with higher accuracy.
arXiv Detail & Related papers (2024-07-04T13:46:30Z) - Chest X-ray Image Classification: A Causal Perspective [49.87607548975686]
We propose a causal approach to address the CXR classification problem, which constructs a structural causal model (SCM) and uses the backdoor adjustment to select effective visual information for CXR classification.
Experimental results demonstrate that our proposed method outperforms the open-source NIH ChestX-ray14 in terms of classification performance.
arXiv Detail & Related papers (2023-05-20T03:17:44Z) - Cross-modulated Few-shot Image Generation for Colorectal Tissue
Classification [58.147396879490124]
Our few-shot generation method, named XM-GAN, takes one base and a pair of reference tissue images as input and generates high-quality yet diverse images.
To the best of our knowledge, we are the first to investigate few-shot generation in colorectal tissue images.
arXiv Detail & Related papers (2023-04-04T17:50:30Z) - Can we Adopt Self-supervised Pretraining for Chest X-Rays? [10.529356372817386]
Convolutional Neural Networks (CNN) have seen success in identifying pathologies in Chest X-Ray (CXR) images.
In this work, we analyze the utility of pretraining on unlabeled ImageNet or CXR datasets using various algorithms and in multiple settings.
Some findings of our work include: (i) supervised training with labeled ImageNet learns strong representations that are hard to beat; (ii) self-supervised pretraining on ImageNet (1M images) shows performance similar to self-supervised pretraining on a CXR dataset (100K images); and (iii) the CNN trained on supervised
arXiv Detail & Related papers (2022-11-23T13:06:59Z) - Optimising Chest X-Rays for Image Analysis by Identifying and Removing
Confounding Factors [49.005337470305584]
During the COVID-19 pandemic, the sheer volume of imaging performed in an emergency setting for COVID-19 diagnosis has resulted in a wide variability of clinical CXR acquisitions.
The variable quality of clinically-acquired CXRs within publicly available datasets could have a profound effect on algorithm performance.
We propose a simple and effective step-wise approach to pre-processing a COVID-19 chest X-ray dataset to remove undesired biases.
arXiv Detail & Related papers (2022-08-22T13:57:04Z) - Radiomics-Guided Global-Local Transformer for Weakly Supervised
Pathology Localization in Chest X-Rays [65.88435151891369]
Radiomics-Guided Transformer (RGT) fuses textitglobal image information with textitlocal knowledge-guided radiomics information.
RGT consists of an image Transformer branch, a radiomics Transformer branch, and fusion layers that aggregate image and radiomic information.
arXiv Detail & Related papers (2022-07-10T06:32:56Z) - Preservation of High Frequency Content for Deep Learning-Based Medical
Image Classification [74.84221280249876]
An efficient analysis of large amounts of chest radiographs can aid physicians and radiologists.
We propose a novel Discrete Wavelet Transform (DWT)-based method for the efficient identification and encoding of visual information.
arXiv Detail & Related papers (2022-05-08T15:29:54Z) - Multi-Label Generalized Zero Shot Learning for the Classification of
Disease in Chest Radiographs [0.7734726150561088]
We propose a zero shot learning network that can simultaneously predict multiple seen and unseen diseases in chest X-ray images.
The network is end-to-end trainable and requires no independent pre-training for the offline feature extractor.
Our network outperforms two strong baselines in terms of recall, precision, f1 score, and area under the receiver operating characteristic curve.
arXiv Detail & Related papers (2021-07-14T09:04:20Z) - Image Embedding and Model Ensembling for Automated Chest X-Ray
Interpretation [0.0]
We present and study several machine learning approaches to develop automated Chest X-ray diagnostic models.
In particular, we trained several Convolutional Neural Networks (CNN) on the CheXpert dataset.
We used the trained CNNs to compute embeddings of the CXR images, in order to train two sets of tree-based classifiers from them.
arXiv Detail & Related papers (2021-05-05T14:48:59Z) - Cross-Modal Contrastive Learning for Abnormality Classification and
Localization in Chest X-rays with Radiomics using a Feedback Loop [63.81818077092879]
We propose an end-to-end semi-supervised cross-modal contrastive learning framework for medical images.
We first apply an image encoder to classify the chest X-rays and to generate the image features.
The radiomic features are then passed through another dedicated encoder to act as the positive sample for the image features generated from the same chest X-ray.
arXiv Detail & Related papers (2021-04-11T09:16:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.