Pyramid Attention Network for Medical Image Registration
- URL: http://arxiv.org/abs/2402.09016v1
- Date: Wed, 14 Feb 2024 08:46:18 GMT
- Title: Pyramid Attention Network for Medical Image Registration
- Authors: Zhuoyuan Wang, Haiqiao Wang, Yi Wang
- Abstract summary: We propose a pyramid attention network (PAN) for deformable medical image registration.
PAN incorporates a dual-stream pyramid encoder with channel-wise attention to boost the feature representation.
Our method achieves favorable registration performance, while outperforming several CNN-based and Transformer-based registration networks.
- Score: 4.142556531859984
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The advent of deep-learning-based registration networks has addressed the
time-consuming challenge in traditional iterative methods.However, the
potential of current registration networks for comprehensively capturing
spatial relationships has not been fully explored, leading to inadequate
performance in large-deformation image registration.The pure convolutional
neural networks (CNNs) neglect feature enhancement, while current
Transformer-based networks are susceptible to information redundancy.To
alleviate these issues, we propose a pyramid attention network (PAN) for
deformable medical image registration.Specifically, the proposed PAN
incorporates a dual-stream pyramid encoder with channel-wise attention to boost
the feature representation.Moreover, a multi-head local attention Transformer
is introduced as decoder to analyze motion patterns and generate deformation
fields.Extensive experiments on two public brain magnetic resonance imaging
(MRI) datasets and one abdominal MRI dataset demonstrate that our method
achieves favorable registration performance, while outperforming several
CNN-based and Transformer-based registration networks.Our code is publicly
available at https://github.com/JuliusWang-7/PAN.
Related papers
- MrRegNet: Multi-resolution Mask Guided Convolutional Neural Network for Medical Image Registration with Large Deformations [6.919880141683284]
MrRegNet is a mask-guided encoder-decoder DCNN-based image registration method.
Image alignment accuracies are significantly improved at local regions guided by segmentation masks.
arXiv Detail & Related papers (2024-05-16T12:57:03Z) - TCCT-Net: Two-Stream Network Architecture for Fast and Efficient Engagement Estimation via Behavioral Feature Signals [58.865901821451295]
We present a novel two-stream feature fusion "Tensor-Convolution and Convolution-Transformer Network" (TCCT-Net) architecture.
To better learn the meaningful patterns in the temporal-spatial domain, we design a "CT" stream that integrates a hybrid convolutional-transformer.
In parallel, to efficiently extract rich patterns from the temporal-frequency domain, we introduce a "TC" stream that uses Continuous Wavelet Transform (CWT) to represent information in a 2D tensor form.
arXiv Detail & Related papers (2024-04-15T06:01:48Z) - TBSN: Transformer-Based Blind-Spot Network for Self-Supervised Image Denoising [94.09442506816724]
Blind-spot networks (BSN) have been prevalent network architectures in self-supervised image denoising (SSID)
We present a transformer-based blind-spot network (TBSN) by analyzing and redesigning the transformer operators that meet the blind-spot requirement.
For spatial self-attention, an elaborate mask is applied to the attention matrix to restrict its receptive field, thus mimicking the dilated convolution.
For channel self-attention, we observe that it may leak the blind-spot information when the channel number is greater than spatial size in the deep layers of multi-scale architectures.
arXiv Detail & Related papers (2024-04-11T15:39:10Z) - Affine-Consistent Transformer for Multi-Class Cell Nuclei Detection [76.11864242047074]
We propose a novel Affine-Consistent Transformer (AC-Former), which directly yields a sequence of nucleus positions.
We introduce an Adaptive Affine Transformer (AAT) module, which can automatically learn the key spatial transformations to warp original images for local network training.
Experimental results demonstrate that the proposed method significantly outperforms existing state-of-the-art algorithms on various benchmarks.
arXiv Detail & Related papers (2023-10-22T02:27:02Z) - RetSeg: Retention-based Colorectal Polyps Segmentation Network [0.0]
Vision Transformers (ViTs) have revolutionized medical imaging analysis.
ViTs exhibit contextual awareness in processing visual data, culminating in robust and precise predictions.
We introduce RetSeg, an encoder-decoder network featuring multi-head retention blocks.
arXiv Detail & Related papers (2023-10-09T06:43:38Z) - ModeT: Learning Deformable Image Registration via Motion Decomposition
Transformer [7.629385629884155]
We propose a novel motion decomposition Transformer (ModeT) to explicitly model multiple motion modalities.
Our method outperforms current state-of-the-art registration networks and Transformers.
arXiv Detail & Related papers (2023-06-09T06:00:05Z) - Focused Decoding Enables 3D Anatomical Detection by Transformers [64.36530874341666]
We propose a novel Detection Transformer for 3D anatomical structure detection, dubbed Focused Decoder.
Focused Decoder leverages information from an anatomical region atlas to simultaneously deploy query anchors and restrict the cross-attention's field of view.
We evaluate our proposed approach on two publicly available CT datasets and demonstrate that Focused Decoder not only provides strong detection results and thus alleviates the need for a vast amount of annotated data but also exhibits exceptional and highly intuitive explainability of results via attention weights.
arXiv Detail & Related papers (2022-07-21T22:17:21Z) - Symmetric Transformer-based Network for Unsupervised Image Registration [4.258536928793156]
We propose a convolution-based efficient multi-head self-attention (CEMSA) block, which reduces the parameters of the traditional Transformer.
Based on the proposed CEMSA, we present a novel Symmetric Transformer-based model (SymTrans)
Experimental results show that our proposed method achieves state-of-the-art performance in image registration.
arXiv Detail & Related papers (2022-04-28T15:45:09Z) - CAR-Net: Unsupervised Co-Attention Guided Registration Network for Joint
Registration and Structure Learning [73.03885837923599]
We propose a novel deep learning network, Co-Attention guided Registration Network (CAR-Net)
CAR-Net employs a co-attention block to learn a new representation of the inputs, which drives the registration of the fixed and moving images.
Experiments on UK Biobank cardiac cine-magnetic resonance image data demonstrate that CAR-Net obtains higher registration accuracy and smoother deformation fields than state-of-the-art unsupervised registration methods.
arXiv Detail & Related papers (2021-06-11T23:25:49Z) - Less is More: Pay Less Attention in Vision Transformers [61.05787583247392]
Less attention vIsion Transformer builds upon the fact that convolutions, fully-connected layers, and self-attentions have almost equivalent mathematical expressions for processing image patch sequences.
The proposed LIT achieves promising performance on image recognition tasks, including image classification, object detection and instance segmentation.
arXiv Detail & Related papers (2021-05-29T05:26:07Z) - Unsupervised MMRegNet based on Spatially Encoded Gradient Information [16.355832135847276]
Multi-modality medical images can provide relevant and complementary anatomical information for a target (organ, tumor or tissue)
It is still challenging to develop a multi-modality registration network due to the lack of robust criteria for network training.
In this work, we propose a multi-modality registration network (MMRegNet), which can jointly register multiple images with different modalities to a target image.
arXiv Detail & Related papers (2021-05-16T09:47:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.