Automatic Diagnosis of Malaria from Thin Blood Smear Images using Deep
Convolutional Neural Network with Multi-Resolution Feature Fusion
- URL: http://arxiv.org/abs/2012.05350v1
- Date: Wed, 9 Dec 2020 22:44:05 GMT
- Title: Automatic Diagnosis of Malaria from Thin Blood Smear Images using Deep
Convolutional Neural Network with Multi-Resolution Feature Fusion
- Authors: Tanvir Mahmud and Shaikh Anowarul Fattah
- Abstract summary: An end-to-end deep learning-based approach is proposed for faster diagnosis of malaria from thin blood smear images.
An efficient, highly scalable deep neural network, named as DilationNet, is proposed that incorporates features from a large spectrum by varying dilation rates of convolutions to extract features from different receptive areas.
A feature fusion scheme is introduced with the proposed DeepFusionNet architecture for jointly optimizing the feature space of these individually trained networks.
Experiments on a publicly available dataset show outstanding performance with accuracy over 99.5% outperforming other state-of-the-art approaches.
- Score: 0.7310043452300736
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Malaria, a life-threatening disease, infects millions of people every year
throughout the world demanding faster diagnosis for proper treatment before any
damages occur. In this paper, an end-to-end deep learning-based approach is
proposed for faster diagnosis of malaria from thin blood smear images by making
efficient optimizations of features extracted from diversified receptive
fields. Firstly, an efficient, highly scalable deep neural network, named as
DilationNet, is proposed that incorporates features from a large spectrum by
varying dilation rates of convolutions to extract features from different
receptive areas. Next, the raw images are resampled to various resolutions to
introduce variations in the receptive fields that are used for independently
optimizing different forms of DilationNet scaled for different resolutions of
images. Afterward, a feature fusion scheme is introduced with the proposed
DeepFusionNet architecture for jointly optimizing the feature space of these
individually trained networks operating on different levels of observations.
All the convolutional layers of various forms of DilationNets that are
optimized to extract spatial features from different resolutions of images are
directly transferred to provide a variegated feature space from any image.
Later, joint optimization of these spatial features is carried out in the
DeepFusionNet to extract the most relevant representation of the sample image.
This scheme offers the opportunity to explore the feature space extensively by
varying the observation level to accurately diagnose the abnormality. Intense
experimentations on a publicly available dataset show outstanding performance
with accuracy over 99.5% outperforming other state-of-the-art approaches.
Related papers
- Adapting Visual-Language Models for Generalizable Anomaly Detection in Medical Images [68.42215385041114]
This paper introduces a novel lightweight multi-level adaptation and comparison framework to repurpose the CLIP model for medical anomaly detection.
Our approach integrates multiple residual adapters into the pre-trained visual encoder, enabling a stepwise enhancement of visual features across different levels.
Our experiments on medical anomaly detection benchmarks demonstrate that our method significantly surpasses current state-of-the-art models.
arXiv Detail & Related papers (2024-03-19T09:28:19Z) - DiAD: A Diffusion-based Framework for Multi-class Anomaly Detection [55.48770333927732]
We propose a Difusion-based Anomaly Detection (DiAD) framework for multi-class anomaly detection.
It consists of a pixel-space autoencoder, a latent-space Semantic-Guided (SG) network with a connection to the stable diffusion's denoising network, and a feature-space pre-trained feature extractor.
Experiments on MVTec-AD and VisA datasets demonstrate the effectiveness of our approach.
arXiv Detail & Related papers (2023-12-11T18:38:28Z) - FeaInfNet: Diagnosis in Medical Image with Feature-Driven Inference and
Visual Explanations [4.022446255159328]
Interpretable deep learning models have received widespread attention in the field of image recognition.
Many interpretability models that have been proposed still have problems of insufficient accuracy and interpretability in medical image disease diagnosis.
We propose feature-driven inference network (FeaInfNet) to solve these problems.
arXiv Detail & Related papers (2023-12-04T13:09:00Z) - EMIT-Diff: Enhancing Medical Image Segmentation via Text-Guided
Diffusion Model [4.057796755073023]
We develop controllable diffusion models for medical image synthesis, called EMIT-Diff.
We leverage recent diffusion probabilistic models to generate realistic and diverse synthetic medical image data.
In our approach, we ensure that the synthesized samples adhere to medically relevant constraints.
arXiv Detail & Related papers (2023-10-19T16:18:02Z) - Multi-modal Medical Neurological Image Fusion using Wavelet Pooled Edge
Preserving Autoencoder [3.3828292731430545]
This paper presents an end-to-end unsupervised fusion model for multimodal medical images based on an edge-preserving dense autoencoder network.
In the proposed model, feature extraction is improved by using wavelet decomposition-based attention pooling of feature maps.
The proposed model is trained on a variety of medical image pairs which helps in capturing the intensity distributions of the source images.
arXiv Detail & Related papers (2023-10-18T11:59:35Z) - VesselMorph: Domain-Generalized Retinal Vessel Segmentation via
Shape-Aware Representation [12.194439938007672]
Domain shift is an inherent property of medical images and has become a major obstacle for large-scale deployment of learning-based algorithms.
We propose a method named VesselMorph which generalizes the 2D retinal vessel segmentation task by synthesizing a shape-aware representation.
VesselMorph achieves superior generalization performance compared with competing methods in different domain shift scenarios.
arXiv Detail & Related papers (2023-07-01T06:02:22Z) - Affinity Feature Strengthening for Accurate, Complete and Robust Vessel
Segmentation [48.638327652506284]
Vessel segmentation is crucial in many medical image applications, such as detecting coronary stenoses, retinal vessel diseases and brain aneurysms.
We present a novel approach, the affinity feature strengthening network (AFN), which jointly models geometry and refines pixel-wise segmentation features using a contrast-insensitive, multiscale affinity approach.
arXiv Detail & Related papers (2022-11-12T05:39:17Z) - RetiFluidNet: A Self-Adaptive and Multi-Attention Deep Convolutional
Network for Retinal OCT Fluid Segmentation [3.57686754209902]
Quantification of retinal fluids is necessary for OCT-guided treatment management.
New convolutional neural architecture named RetiFluidNet is proposed for multi-class retinal fluid segmentation.
Model benefits from hierarchical representation learning of textural, contextual, and edge features.
arXiv Detail & Related papers (2022-09-26T07:18:00Z) - PC-GANs: Progressive Compensation Generative Adversarial Networks for
Pan-sharpening [50.943080184828524]
We propose a novel two-step model for pan-sharpening that sharpens the MS image through the progressive compensation of the spatial and spectral information.
The whole model is composed of triple GANs, and based on the specific architecture, a joint compensation loss function is designed to enable the triple GANs to be trained simultaneously.
arXiv Detail & Related papers (2022-07-29T03:09:21Z) - Improved Slice-wise Tumour Detection in Brain MRIs by Computing
Dissimilarities between Latent Representations [68.8204255655161]
Anomaly detection for Magnetic Resonance Images (MRIs) can be solved with unsupervised methods.
We have proposed a slice-wise semi-supervised method for tumour detection based on the computation of a dissimilarity function in the latent space of a Variational AutoEncoder.
We show that by training the models on higher resolution images and by improving the quality of the reconstructions, we obtain results which are comparable with different baselines.
arXiv Detail & Related papers (2020-07-24T14:02:09Z) - Pathological Retinal Region Segmentation From OCT Images Using Geometric
Relation Based Augmentation [84.7571086566595]
We propose improvements over previous GAN-based medical image synthesis methods by jointly encoding the intrinsic relationship of geometry and shape.
The proposed method outperforms state-of-the-art segmentation methods on the public RETOUCH dataset having images captured from different acquisition procedures.
arXiv Detail & Related papers (2020-03-31T11:50:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.