Enhancing Retinal Vascular Structure Segmentation in Images With a Novel
Design Two-Path Interactive Fusion Module Model
- URL: http://arxiv.org/abs/2403.01362v1
- Date: Sun, 3 Mar 2024 01:36:11 GMT
- Title: Enhancing Retinal Vascular Structure Segmentation in Images With a Novel
Design Two-Path Interactive Fusion Module Model
- Authors: Rui Yang and Shunpu Zhang
- Abstract summary: We introduce Swin-Res-Net, a specialized module designed to enhance the precision of retinal vessel segmentation.
Swin-Res-Net utilizes the Swin transformer which uses shifted windows with displacement for partitioning.
Our proposed architecture produces outstanding results, either meeting or surpassing those of other published models.
- Score: 6.392575673488379
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Precision in identifying and differentiating micro and macro blood vessels in
the retina is crucial for the diagnosis of retinal diseases, although it poses
a significant challenge. Current autoencoding-based segmentation approaches
encounter limitations as they are constrained by the encoder and undergo a
reduction in resolution during the encoding stage. The inability to recover
lost information in the decoding phase further impedes these approaches.
Consequently, their capacity to extract the retinal microvascular structure is
restricted. To address this issue, we introduce Swin-Res-Net, a specialized
module designed to enhance the precision of retinal vessel segmentation.
Swin-Res-Net utilizes the Swin transformer which uses shifted windows with
displacement for partitioning, to reduce network complexity and accelerate
model convergence. Additionally, the model incorporates interactive fusion with
a functional module in the Res2Net architecture. The Res2Net leverages
multi-scale techniques to enlarge the receptive field of the convolutional
kernel, enabling the extraction of additional semantic information from the
image. This combination creates a new module that enhances the localization and
separation of micro vessels in the retina. To improve the efficiency of
processing vascular information, we've added a module to eliminate redundant
information between the encoding and decoding steps.
Our proposed architecture produces outstanding results, either meeting or
surpassing those of other published models. The AUC reflects significant
enhancements, achieving values of 0.9956, 0.9931, and 0.9946 in pixel-wise
segmentation of retinal vessels across three widely utilized datasets:
CHASE-DB1, DRIVE, and STARE, respectively. Moreover, Swin-Res-Net outperforms
alternative architectures, demonstrating superior performance in both IOU and
F1 measure metrics.
Related papers
- TransUNext: towards a more advanced U-shaped framework for automatic vessel segmentation in the fundus image [19.16680702780529]
We propose a more advanced U-shaped architecture for a hybrid Transformer and CNN: TransUNext.
The Global Multi-Scale Fusion (GMSF) module is further introduced to upgrade skip-connections, fuse high-level semantic and low-level detailed information, and eliminate high- and low-level semantic differences.
arXiv Detail & Related papers (2024-11-05T01:44:22Z) - Dual-scale Enhanced and Cross-generative Consistency Learning for Semi-supervised Medical Image Segmentation [49.57907601086494]
Medical image segmentation plays a crucial role in computer-aided diagnosis.
We propose a novel Dual-scale Enhanced and Cross-generative consistency learning framework for semi-supervised medical image (DEC-Seg)
arXiv Detail & Related papers (2023-12-26T12:56:31Z) - Affine-Consistent Transformer for Multi-Class Cell Nuclei Detection [76.11864242047074]
We propose a novel Affine-Consistent Transformer (AC-Former), which directly yields a sequence of nucleus positions.
We introduce an Adaptive Affine Transformer (AAT) module, which can automatically learn the key spatial transformations to warp original images for local network training.
Experimental results demonstrate that the proposed method significantly outperforms existing state-of-the-art algorithms on various benchmarks.
arXiv Detail & Related papers (2023-10-22T02:27:02Z) - Mutual Information-driven Triple Interaction Network for Efficient Image
Dehazing [54.168567276280505]
We propose a novel Mutual Information-driven Triple interaction Network (MITNet) for image dehazing.
The first stage, named amplitude-guided haze removal, aims to recover the amplitude spectrum of the hazy images for haze removal.
The second stage, named phase-guided structure refined, devotes to learning the transformation and refinement of the phase spectrum.
arXiv Detail & Related papers (2023-08-14T08:23:58Z) - Scale-aware Super-resolution Network with Dual Affinity Learning for
Lesion Segmentation from Medical Images [50.76668288066681]
We present a scale-aware super-resolution network to adaptively segment lesions of various sizes from low-resolution medical images.
Our proposed network achieved consistent improvements compared to other state-of-the-art methods.
arXiv Detail & Related papers (2023-05-30T14:25:55Z) - DualStreamFoveaNet: A Dual Stream Fusion Architecture with Anatomical Awareness for Robust Fovea Localization [6.278444803136043]
We propose a novel transformer-based architecture called DualStreamFoveaNet (DSFN) for multi-cue fusion.
This architecture explicitly incorporates long-range connections and global features using retina and vessel distributions for robust fovea localization.
We demonstrate that the DSFN is more robust on both normal and diseased retina images and has better capacity generalization in cross-dataset experiments.
arXiv Detail & Related papers (2023-02-14T10:40:20Z) - DoubleU-NetPlus: A Novel Attention and Context Guided Dual U-Net with
Multi-Scale Residual Feature Fusion Network for Semantic Segmentation of
Medical Images [2.20200533591633]
We present a novel dual U-Net-based architecture named DoubleU-NetPlus.
We exploit multi-contextual features and several attention strategies to increase networks' ability to model discriminative feature representation.
To mitigate the gradient vanishing issue and incorporate high-resolution features with deeper spatial details, the standard convolution operation is replaced with the attention-guided residual convolution operations.
arXiv Detail & Related papers (2022-11-25T16:56:26Z) - RetiFluidNet: A Self-Adaptive and Multi-Attention Deep Convolutional
Network for Retinal OCT Fluid Segmentation [3.57686754209902]
Quantification of retinal fluids is necessary for OCT-guided treatment management.
New convolutional neural architecture named RetiFluidNet is proposed for multi-class retinal fluid segmentation.
Model benefits from hierarchical representation learning of textural, contextual, and edge features.
arXiv Detail & Related papers (2022-09-26T07:18:00Z) - RV-GAN : Retinal Vessel Segmentation from Fundus Images using
Multi-scale Generative Adversarial Networks [0.0]
RVGAN is a new multi-scale generative architecture for accurate retinal vessel segmentation.
Our architecture uses two generators and two multi-scale autoencoder based discriminators, for better microvessel localization and segmentation.
arXiv Detail & Related papers (2021-01-03T01:04:49Z) - Rethinking the Extraction and Interaction of Multi-Scale Features for
Vessel Segmentation [53.187152856583396]
We propose a novel deep learning model called PC-Net to segment retinal vessels and major arteries in 2D fundus image and 3D computed tomography angiography (CTA) scans.
In PC-Net, the pyramid squeeze-and-excitation (PSE) module introduces spatial information to each convolutional block, boosting its ability to extract more effective multi-scale features.
arXiv Detail & Related papers (2020-10-09T08:22:54Z) - DONet: Dual Objective Networks for Skin Lesion Segmentation [77.9806410198298]
We propose a simple yet effective framework, named Dual Objective Networks (DONet), to improve the skin lesion segmentation.
Our DONet adopts two symmetric decoders to produce different predictions for approaching different objectives.
To address the challenge of large variety of lesion scales and shapes in dermoscopic images, we additionally propose a recurrent context encoding module (RCEM)
arXiv Detail & Related papers (2020-08-19T06:02:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.