DualAttNet: Synergistic Fusion of Image-level and Fine-Grained Disease
Attention for Multi-Label Lesion Detection in Chest X-rays
- URL: http://arxiv.org/abs/2306.13813v1
- Date: Fri, 23 Jun 2023 23:19:27 GMT
- Title: DualAttNet: Synergistic Fusion of Image-level and Fine-Grained Disease
Attention for Multi-Label Lesion Detection in Chest X-rays
- Authors: Qing Xu and Wenting Duan
- Abstract summary: We propose a dual attention supervised module for multi-label lesion detection in chest radiographs, named DualAttNet.
It efficiently fuses global and local lesion classification information based on an image-level attention block and a fine-grained disease attention algorithm.
- Score: 1.3367903535457364
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Chest radiographs are the most commonly performed radiological examinations
for lesion detection. Recent advances in deep learning have led to encouraging
results in various thoracic disease detection tasks. Particularly, the
architecture with feature pyramid network performs the ability to recognise
targets with different sizes. However, such networks are difficult to focus on
lesion regions in chest X-rays due to their high resemblance in vision. In this
paper, we propose a dual attention supervised module for multi-label lesion
detection in chest radiographs, named DualAttNet. It efficiently fuses global
and local lesion classification information based on an image-level attention
block and a fine-grained disease attention algorithm. A binary cross entropy
loss function is used to calculate the difference between the attention map and
ground truth at image level. The generated gradient flow is leveraged to refine
pyramid representations and highlight lesion-related features. We evaluate the
proposed model on VinDr-CXR, ChestX-ray8 and COVID-19 datasets. The
experimental results show that DualAttNet surpasses baselines by 0.6% to 2.7%
mAP and 1.4% to 4.7% AP50 with different detection architectures. The code for
our work and more technical details can be found at
https://github.com/xq141839/DualAttNet.
Related papers
- MultiFusionNet: Multilayer Multimodal Fusion of Deep Neural Networks for
Chest X-Ray Image Classification [16.479941416339265]
Automated systems utilizing convolutional neural networks (CNNs) have shown promise in improving the accuracy and efficiency of chest X-ray image classification.
We propose a novel deep learning-based multilayer multimodal fusion model that emphasizes extracting features from different layers and fusing them.
The proposed model achieves a significantly higher accuracy of 97.21% and 99.60% for both three-class and two-class classifications, respectively.
arXiv Detail & Related papers (2024-01-01T11:50:01Z) - Multi-Resolution Fusion for Fully Automatic Cephalometric Landmark
Detection [1.9580473532948401]
Cephalometric landmark detection on lateral skull X-ray images plays a crucial role in the diagnosis of certain dental diseases.
Based on extensive data observations and quantitative analyses, we discovered that visual features from different receptive fields affect the detection accuracy of various landmarks differently.
We implemented this method in the Cephalometric Landmark Detection in Lateral X-ray Images 2023 Challenge and achieved a Mean Radial Error (MRE) of 1.62 mm and a Success Detection Rate (SDR) 2.0mm of 74.18% in the final testing phase.
arXiv Detail & Related papers (2023-10-04T14:42:45Z) - EchoGLAD: Hierarchical Graph Neural Networks for Left Ventricle Landmark
Detection on Echocardiograms [21.537833605289816]
We introduce an echocardiogram-based, hierarchical graph neural network (GNN) for left ventricle landmark detection (EchoGLAD)
Our main contributions are 1) a hierarchical graph representation learning framework for multi-resolution landmark detection via GNNs; 2) induced hierarchical supervision at different levels of granularity using a multi-level loss.
arXiv Detail & Related papers (2023-07-23T05:31:47Z) - Data-Efficient Vision Transformers for Multi-Label Disease
Classification on Chest Radiographs [55.78588835407174]
Vision Transformers (ViTs) have not been applied to this task despite their high classification performance on generic images.
ViTs do not rely on convolutions but on patch-based self-attention and in contrast to CNNs, no prior knowledge of local connectivity is present.
Our results show that while the performance between ViTs and CNNs is on par with a small benefit for ViTs, DeiTs outperform the former if a reasonably large data set is available for training.
arXiv Detail & Related papers (2022-08-17T09:07:45Z) - Radiomics-Guided Global-Local Transformer for Weakly Supervised
Pathology Localization in Chest X-Rays [65.88435151891369]
Radiomics-Guided Transformer (RGT) fuses textitglobal image information with textitlocal knowledge-guided radiomics information.
RGT consists of an image Transformer branch, a radiomics Transformer branch, and fusion layers that aggregate image and radiomic information.
arXiv Detail & Related papers (2022-07-10T06:32:56Z) - Preservation of High Frequency Content for Deep Learning-Based Medical
Image Classification [74.84221280249876]
An efficient analysis of large amounts of chest radiographs can aid physicians and radiologists.
We propose a novel Discrete Wavelet Transform (DWT)-based method for the efficient identification and encoding of visual information.
arXiv Detail & Related papers (2022-05-08T15:29:54Z) - A Deep Learning Approach for the Detection of COVID-19 from Chest X-Ray
Images using Convolutional Neural Networks [0.0]
COVID-19 (coronavirus) is an ongoing pandemic caused by severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2)
The virus was first identified in mid-December 2019 in the Hubei province of Wuhan, China.
It has spread throughout the planet with more than 75.5 million confirmed cases and more than 1.67 million deaths.
arXiv Detail & Related papers (2022-01-24T21:12:25Z) - Cross-Modal Contrastive Learning for Abnormality Classification and
Localization in Chest X-rays with Radiomics using a Feedback Loop [63.81818077092879]
We propose an end-to-end semi-supervised cross-modal contrastive learning framework for medical images.
We first apply an image encoder to classify the chest X-rays and to generate the image features.
The radiomic features are then passed through another dedicated encoder to act as the positive sample for the image features generated from the same chest X-ray.
arXiv Detail & Related papers (2021-04-11T09:16:29Z) - Many-to-One Distribution Learning and K-Nearest Neighbor Smoothing for
Thoracic Disease Identification [83.6017225363714]
deep learning has become the most powerful computer-aided diagnosis technology for improving disease identification performance.
For chest X-ray imaging, annotating large-scale data requires professional domain knowledge and is time-consuming.
In this paper, we propose many-to-one distribution learning (MODL) and K-nearest neighbor smoothing (KNNS) methods to improve a single model's disease identification performance.
arXiv Detail & Related papers (2021-02-26T02:29:30Z) - Bifurcated Autoencoder for Segmentation of COVID-19 Infected Regions in
CT Images [8.122848195290743]
The new coronavirus infection COVID-19 has shocked the world since early 2020 with its aggressive outbreak.
Deep learning and convolutional neural networks have been used for image analysis in this context.
This paper proposes an approach to segment lung regions infected by COVID-19 to help cardiologists diagnose the disease more accurately, faster, and more manageable.
arXiv Detail & Related papers (2020-11-01T21:32:00Z) - ElixirNet: Relation-aware Network Architecture Adaptation for Medical
Lesion Detection [90.13718478362337]
We introduce a novel ElixirNet that includes three components: 1) TruncatedRPN balances positive and negative data for false positive reduction; 2) Auto-lesion Block is automatically customized for medical images to incorporate relation-aware operations among region proposals; and 3) Relation transfer module incorporates the semantic relationship.
Experiments on DeepLesion and Kits19 prove the effectiveness of ElixirNet, achieving improvement of both sensitivity and precision over FPN with fewer parameters.
arXiv Detail & Related papers (2020-03-03T05:29:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.