PAENet: A Progressive Attention-Enhanced Network for 3D to 2D Retinal
Vessel Segmentation
- URL: http://arxiv.org/abs/2108.11695v1
- Date: Thu, 26 Aug 2021 10:27:25 GMT
- Title: PAENet: A Progressive Attention-Enhanced Network for 3D to 2D Retinal
Vessel Segmentation
- Authors: Zhuojie Wu and Muyi Sun
- Abstract summary: 3D to 2D retinal vessel segmentation is a challenging problem in Optical Coherence Tomography Angiography ( OCTA) images.
We propose a Progressive Attention-Enhanced Network (PAENet) based on attention mechanisms to extract rich feature representation.
Our proposed algorithm achieves state-of-the-art performance compared with previous methods.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: 3D to 2D retinal vessel segmentation is a challenging problem in Optical
Coherence Tomography Angiography (OCTA) images. Accurate retinal vessel
segmentation is important for the diagnosis and prevention of ophthalmic
diseases. However, making full use of the 3D data of OCTA volumes is a vital
factor for obtaining satisfactory segmentation results. In this paper, we
propose a Progressive Attention-Enhanced Network (PAENet) based on attention
mechanisms to extract rich feature representation. Specifically, the framework
consists of two main parts, the three-dimensional feature learning path and the
two-dimensional segmentation path. In the three-dimensional feature learning
path, we design a novel Adaptive Pooling Module (APM) and propose a new
Quadruple Attention Module (QAM). The APM captures dependencies along the
projection direction of volumes and learns a series of pooling coefficients for
feature fusion, which efficiently reduces feature dimension. In addition, the
QAM reweights the features by capturing four-group cross-dimension
dependencies, which makes maximum use of 4D feature tensors. In the
two-dimensional segmentation path, to acquire more detailed information, we
propose a Feature Fusion Module (FFM) to inject 3D information into the 2D
path. Meanwhile, we adopt the Polarized Self-Attention (PSA) block to model the
semantic interdependencies in spatial and channel dimensions respectively.
Experimentally, our extensive experiments on the OCTA-500 dataset show that our
proposed algorithm achieves state-of-the-art performance compared with previous
methods.
Related papers
- 4D Contrastive Superflows are Dense 3D Representation Learners [62.433137130087445]
We introduce SuperFlow, a novel framework designed to harness consecutive LiDAR-camera pairs for establishing pretraining objectives.
To further boost learning efficiency, we incorporate a plug-and-play view consistency module that enhances alignment of the knowledge distilled from camera views.
arXiv Detail & Related papers (2024-07-08T17:59:54Z) - A Flexible 2.5D Medical Image Segmentation Approach with In-Slice and Cross-Slice Attention [13.895277069418045]
We introduce CSA-Net, a flexible 2.5D segmentation model capable of processing 2.5D images with an arbitrary number of slices.
We evaluate CSA-Net on three 2.5D segmentation tasks: (1) brain MRI segmentation, (2) binary prostate MRI segmentation, and (3) multi-class prostate MRI segmentation.
arXiv Detail & Related papers (2024-04-30T18:28:09Z) - 3D Vascular Segmentation Supervised by 2D Annotation of Maximum
Intensity Projection [33.34240545722551]
Vascular structure segmentation plays a crucial role in medical analysis and clinical applications.
Existing weakly supervised methods have exhibited suboptimal performance when handling sparse vascular structure.
Here, we employ maximum intensity projection (MIP) to decrease the dimensionality of 3D volume to 2D image for efficient annotation.
We introduce a weakly-supervised network that fuses 2D-3D deep features via MIP to further improve segmentation performance.
arXiv Detail & Related papers (2024-02-19T13:24:46Z) - Spatiotemporal Modeling Encounters 3D Medical Image Analysis:
Slice-Shift UNet with Multi-View Fusion [0.0]
We propose a new 2D-based model dubbed Slice SHift UNet which encodes three-dimensional features at 2D CNN's complexity.
More precisely multi-view features are collaboratively learned by performing 2D convolutions along the three planes of a volume.
The effectiveness of our approach is validated in Multi-Modality Abdominal Multi-Organ axis (AMOS) and Multi-Atlas Labeling Beyond the Cranial Vault (BTCV) datasets.
arXiv Detail & Related papers (2023-07-24T14:53:23Z) - UNETR++: Delving into Efficient and Accurate 3D Medical Image Segmentation [93.88170217725805]
We propose a 3D medical image segmentation approach, named UNETR++, that offers both high-quality segmentation masks as well as efficiency in terms of parameters, compute cost, and inference speed.
The core of our design is the introduction of a novel efficient paired attention (EPA) block that efficiently learns spatial and channel-wise discriminative features.
Our evaluations on five benchmarks, Synapse, BTCV, ACDC, BRaTs, and Decathlon-Lung, reveal the effectiveness of our contributions in terms of both efficiency and accuracy.
arXiv Detail & Related papers (2022-12-08T18:59:57Z) - Conquering Data Variations in Resolution: A Slice-Aware Multi-Branch
Decoder Network [28.946037652152395]
We identify the wide variation in the ratio between intra- and inter-slice resolutions as a crucial obstacle to the performance.
We propose a slice-aware 2.5D network that emphasizes extracting discnative features utilizing not only in-plane semantics but also out-of-plane for each separate slice.
arXiv Detail & Related papers (2022-03-07T14:31:26Z) - Lymphoma segmentation from 3D PET-CT images using a deep evidential
network [20.65641432056608]
An automatic evidential segmentation method is proposed to segment lymphomas from 3D Positron Emission Tomography (PET) and Computed Tomography (CT) images.
The architecture is composed of a deep feature-extraction module and an evidential layer.
The proposed combination of deep feature extraction and evidential segmentation is shown to outperform the baseline UNet model.
arXiv Detail & Related papers (2022-01-31T09:34:38Z) - Cylindrical and Asymmetrical 3D Convolution Networks for LiDAR-based
Perception [122.53774221136193]
State-of-the-art methods for driving-scene LiDAR-based perception often project the point clouds to 2D space and then process them via 2D convolution.
A natural remedy is to utilize the 3D voxelization and 3D convolution network.
We propose a new framework for the outdoor LiDAR segmentation, where cylindrical partition and asymmetrical 3D convolution networks are designed to explore the 3D geometric pattern.
arXiv Detail & Related papers (2021-09-12T06:25:11Z) - FS-Net: Fast Shape-based Network for Category-Level 6D Object Pose
Estimation with Decoupled Rotation Mechanism [49.89268018642999]
We propose a fast shape-based network (FS-Net) with efficient category-level feature extraction for 6D pose estimation.
The proposed method achieves state-of-the-art performance in both category- and instance-level 6D object pose estimation.
arXiv Detail & Related papers (2021-03-12T03:07:24Z) - Revisiting 3D Context Modeling with Supervised Pre-training for
Universal Lesion Detection in CT Slices [48.85784310158493]
We propose a Modified Pseudo-3D Feature Pyramid Network (MP3D FPN) to efficiently extract 3D context enhanced 2D features for universal lesion detection in CT slices.
With the novel pre-training method, the proposed MP3D FPN achieves state-of-the-art detection performance on the DeepLesion dataset.
The proposed 3D pre-trained weights can potentially be used to boost the performance of other 3D medical image analysis tasks.
arXiv Detail & Related papers (2020-12-16T07:11:16Z) - Improving Point Cloud Semantic Segmentation by Learning 3D Object
Detection [102.62963605429508]
Point cloud semantic segmentation plays an essential role in autonomous driving.
Current 3D semantic segmentation networks focus on convolutional architectures that perform great for well represented classes.
We propose a novel Aware 3D Semantic Detection (DASS) framework that explicitly leverages localization features from an auxiliary 3D object detection task.
arXiv Detail & Related papers (2020-09-22T14:17:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.