Preservation of High Frequency Content for Deep Learning-Based Medical
Image Classification
- URL: http://arxiv.org/abs/2205.03898v1
- Date: Sun, 8 May 2022 15:29:54 GMT
- Title: Preservation of High Frequency Content for Deep Learning-Based Medical
Image Classification
- Authors: Declan McIntosh and Tunai Porto Marques and Alexandra Branzan Albu
- Abstract summary: An efficient analysis of large amounts of chest radiographs can aid physicians and radiologists.
We propose a novel Discrete Wavelet Transform (DWT)-based method for the efficient identification and encoding of visual information.
- Score: 74.84221280249876
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Chest radiographs are used for the diagnosis of multiple critical illnesses
(e.g., Pneumonia, heart failure, lung cancer), for this reason, systems for the
automatic or semi-automatic analysis of these data are of particular interest.
An efficient analysis of large amounts of chest radiographs can aid physicians
and radiologists, ultimately allowing for better medical care of lung-, heart-
and chest-related conditions. We propose a novel Discrete Wavelet Transform
(DWT)-based method for the efficient identification and encoding of visual
information that is typically lost in the down-sampling of high-resolution
radiographs, a common step in computer-aided diagnostic pipelines. Our proposed
approach requires only slight modifications to the input of existing
state-of-the-art Convolutional Neural Networks (CNNs), making it easily
applicable to existing image classification frameworks. We show that the extra
high-frequency components offered by our method increased the classification
performance of several CNNs in benchmarks employing the NIH Chest-8 and
ImageNet-2017 datasets. Based on our results we hypothesize that providing
frequency-specific coefficients allows the CNNs to specialize in the
identification of structures that are particular to a frequency band,
ultimately increasing classification performance, without an increase in
computational load. The implementation of our work is available at
github.com/DeclanMcIntosh/LeGallCuda.
Related papers
- Explanations of Classifiers Enhance Medical Image Segmentation via
End-to-end Pre-training [37.11542605885003]
Medical image segmentation aims to identify and locate abnormal structures in medical images, such as chest radiographs, using deep neural networks.
Our work collects explanations from well-trained classifiers to generate pseudo labels of segmentation tasks.
We then use Integrated Gradients (IG) method to distill and boost the explanations obtained from the classifiers, generating massive diagnosis-oriented localization labels (DoLL)
These DoLL-annotated images are used for pre-training the model before fine-tuning it for downstream segmentation tasks, including COVID-19 infectious areas, lungs, heart, and clavicles.
arXiv Detail & Related papers (2024-01-16T16:18:42Z) - BS-Diff: Effective Bone Suppression Using Conditional Diffusion Models
from Chest X-Ray Images [21.19843479423806]
Chest X-rays (CXRs) are commonly utilized as a low-dose modality for lung screening.
Approximately 75% of the lung area overlaps with bone, which in turn hampers the detection and diagnosis of diseases.
Bone suppression techniques have been introduced, but the current dual-energy subtraction imaging technique in the clinic requires costly equipment and subjects being exposed to high radiation.
This paper proposes a new bone suppression framework, termed BS-Diff, that comprises a conditional diffusion model equipped with a U-Net architecture and a simple enhancement module to incorporate an autoencoder.
arXiv Detail & Related papers (2023-11-26T15:13:13Z) - Swin-Tempo: Temporal-Aware Lung Nodule Detection in CT Scans as Video
Sequences Using Swin Transformer-Enhanced UNet [2.7547288571938795]
We present an innovative model that harnesses the strengths of both convolutional neural networks and vision transformers.
Inspired by object detection in videos, we treat each 3D CT image as a video, individual slices as frames, and lung nodules as objects, enabling a time-series application.
arXiv Detail & Related papers (2023-10-05T07:48:55Z) - High-Fidelity Image Synthesis from Pulmonary Nodule Lesion Maps using
Semantic Diffusion Model [10.412300404240751]
Lung cancer has been one of the leading causes of cancer-related deaths worldwide for years.
Deep learning, computer-assisted diagnosis (CAD) models based on learning algorithms can accelerate the screening process.
However, developing robust and accurate models often requires large-scale and diverse medical datasets with high-quality annotations.
arXiv Detail & Related papers (2023-05-02T01:04:22Z) - Structure Regularized Attentive Network for Automatic Femoral Head
Necrosis Diagnosis and Localization [12.95252724282746]
We propose the structure regularized network attentive (SRANet) to highlight the necrotic regions during classification based on patch attention.
SRANet extracts features in chunks of images, obtains weight via the attention mechanism to aggregate the features, and constrains them by a structural regularizer with prior knowledge to improve the generalization.
Experimental results show that SRANet is superior to CNNs for AVNFH classification, moreover, it can localize lesions and provide more information to assist doctors in diagnosis.
arXiv Detail & Related papers (2022-08-23T02:31:38Z) - Data-Efficient Vision Transformers for Multi-Label Disease
Classification on Chest Radiographs [55.78588835407174]
Vision Transformers (ViTs) have not been applied to this task despite their high classification performance on generic images.
ViTs do not rely on convolutions but on patch-based self-attention and in contrast to CNNs, no prior knowledge of local connectivity is present.
Our results show that while the performance between ViTs and CNNs is on par with a small benefit for ViTs, DeiTs outperform the former if a reasonably large data set is available for training.
arXiv Detail & Related papers (2022-08-17T09:07:45Z) - Generative Residual Attention Network for Disease Detection [51.60842580044539]
We present a novel approach for disease generation in X-rays using a conditional generative adversarial learning.
We generate a corresponding radiology image in a target domain while preserving the identity of the patient.
We then use the generated X-ray image in the target domain to augment our training to improve the detection performance.
arXiv Detail & Related papers (2021-10-25T14:15:57Z) - Many-to-One Distribution Learning and K-Nearest Neighbor Smoothing for
Thoracic Disease Identification [83.6017225363714]
deep learning has become the most powerful computer-aided diagnosis technology for improving disease identification performance.
For chest X-ray imaging, annotating large-scale data requires professional domain knowledge and is time-consuming.
In this paper, we propose many-to-one distribution learning (MODL) and K-nearest neighbor smoothing (KNNS) methods to improve a single model's disease identification performance.
arXiv Detail & Related papers (2021-02-26T02:29:30Z) - Improved Slice-wise Tumour Detection in Brain MRIs by Computing
Dissimilarities between Latent Representations [68.8204255655161]
Anomaly detection for Magnetic Resonance Images (MRIs) can be solved with unsupervised methods.
We have proposed a slice-wise semi-supervised method for tumour detection based on the computation of a dissimilarity function in the latent space of a Variational AutoEncoder.
We show that by training the models on higher resolution images and by improving the quality of the reconstructions, we obtain results which are comparable with different baselines.
arXiv Detail & Related papers (2020-07-24T14:02:09Z) - Y-Net for Chest X-Ray Preprocessing: Simultaneous Classification of
Geometry and Segmentation of Annotations [70.0118756144807]
This work introduces a general pre-processing step for chest x-ray input into machine learning algorithms.
A modified Y-Net architecture based on the VGG11 encoder is used to simultaneously learn geometric orientation and segmentation of radiographs.
Results were evaluated by expert clinicians, with acceptable geometry in 95.8% and annotation mask in 96.2%, compared to 27.0% and 34.9% respectively in control images.
arXiv Detail & Related papers (2020-05-08T02:16:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.