Cross-modal Attention for MRI and Ultrasound Volume Registration
- URL: http://arxiv.org/abs/2107.04548v2
- Date: Mon, 12 Jul 2021 01:46:43 GMT
- Title: Cross-modal Attention for MRI and Ultrasound Volume Registration
- Authors: Xinrui Song, Hengtao Guo, Xuanang Xu, Hanqing Chao, Sheng Xu, Baris
Turkbey, Bradford J. Wood, Ge Wang, Pingkun Yan
- Abstract summary: We develop a self-attention mechanism specifically for cross-modal image registration.
Our proposed cross-modal attention block effectively maps each of the features in one volume to all features in the corresponding volume.
Our experimental results demonstrate that a CNN network designed with the cross-modal attention block embedded outperforms an advanced CNN network 10 times of its size.
- Score: 10.725645523967904
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Prostate cancer biopsy benefits from accurate fusion of transrectal
ultrasound (TRUS) and magnetic resonance (MR) images. In the past few years,
convolutional neural networks (CNNs) have been proved powerful in extracting
image features crucial for image registration. However, challenging
applications and recent advances in computer vision suggest that CNNs are quite
limited in its ability to understand spatial correspondence between features, a
task in which the self-attention mechanism excels. This paper aims to develop a
self-attention mechanism specifically for cross-modal image registration. Our
proposed cross-modal attention block effectively maps each of the features in
one volume to all features in the corresponding volume. Our experimental
results demonstrate that a CNN network designed with the cross-modal attention
block embedded outperforms an advanced CNN network 10 times of its size. We
also incorporated visualization techniques to improve the interpretability of
our network. The source code of our work is available at
https://github.com/DIAL-RPI/Attention-Reg .
Related papers
- Attention Incorporated Network for Sharing Low-rank, Image and K-space Information during MR Image Reconstruction to Achieve Single Breath-hold Cardiac Cine Imaging [9.531827741901662]
We propose to embed information from multiple domains, including low-rank, image, and k-space, in a novel deep learning network for MRI reconstruction.
A-LIKNet adopts a parallel-branch structure, enabling independent learning in the k-space and image domain.
arXiv Detail & Related papers (2024-07-03T11:54:43Z) - Disruptive Autoencoders: Leveraging Low-level features for 3D Medical
Image Pre-training [51.16994853817024]
This work focuses on designing an effective pre-training framework for 3D radiology images.
We introduce Disruptive Autoencoders, a pre-training framework that attempts to reconstruct the original image from disruptions created by a combination of local masking and low-level perturbations.
The proposed pre-training framework is tested across multiple downstream tasks and achieves state-of-the-art performance.
arXiv Detail & Related papers (2023-07-31T17:59:42Z) - Spherical CNN for Medical Imaging Applications: Importance of
Equivariance in image reconstruction and denoising [0.0]
equivariant networks are efficient and high-performance approaches for tomography applications.
We evaluate the efficacy of equivariant spherical CNNs for 2- and 3- dimensional medical imaging problems.
We propose a novel approach to employ SCNNs as a complement to conventional image reconstruction tools.
arXiv Detail & Related papers (2023-07-06T21:18:47Z) - Convolutional neural network based on sparse graph attention mechanism
for MRI super-resolution [0.34410212782758043]
Medical image super-resolution (SR) reconstruction using deep learning techniques can enhance lesion analysis and assist doctors in improving diagnostic efficiency and accuracy.
Existing deep learning-based SR methods rely on convolutional neural networks (CNNs), which inherently limit the expressive capabilities of these models.
We propose an A-network that utilizes multiple convolution operator feature extraction modules (MCO) for extracting image features.
arXiv Detail & Related papers (2023-05-29T06:14:22Z) - Attentive Symmetric Autoencoder for Brain MRI Segmentation [56.02577247523737]
We propose a novel Attentive Symmetric Auto-encoder based on Vision Transformer (ViT) for 3D brain MRI segmentation tasks.
In the pre-training stage, the proposed auto-encoder pays more attention to reconstruct the informative patches according to the gradient metrics.
Experimental results show that our proposed attentive symmetric auto-encoder outperforms the state-of-the-art self-supervised learning methods and medical image segmentation models.
arXiv Detail & Related papers (2022-09-19T09:43:19Z) - Preservation of High Frequency Content for Deep Learning-Based Medical
Image Classification [74.84221280249876]
An efficient analysis of large amounts of chest radiographs can aid physicians and radiologists.
We propose a novel Discrete Wavelet Transform (DWT)-based method for the efficient identification and encoding of visual information.
arXiv Detail & Related papers (2022-05-08T15:29:54Z) - Visual Attention Network [90.0753726786985]
We propose a novel large kernel attention (LKA) module to enable self-adaptive and long-range correlations in self-attention.
We also introduce a novel neural network based on LKA, namely Visual Attention Network (VAN)
VAN outperforms the state-of-the-art vision transformers and convolutional neural networks with a large margin in extensive experiments.
arXiv Detail & Related papers (2022-02-20T06:35:18Z) - Convolutional Neural Network with Convolutional Block Attention Module
for Finger Vein Recognition [4.035753155957698]
We propose a lightweight convolutional neural network with a convolutional block attention module (CBAM) for finger vein recognition.
The experiments are carried out on two publicly available databases and the results demonstrate that the proposed method achieves a stable, highly accurate, and robust performance in multimodal finger recognition.
arXiv Detail & Related papers (2022-02-14T12:59:23Z) - CNNs and GANs in MRI-based cross-modality medical image estimation [1.5469452301122177]
Cross-modality image estimation involves the generation of images of one medical imaging modality from that of another modality.
CNNs have been shown to be useful in identifying, characterising and extracting image patterns.
Generative adversarial networks (GANs) use CNNs as generators and estimated images are discriminated as true or false based on an additional network.
arXiv Detail & Related papers (2021-06-04T01:27:57Z) - The Mind's Eye: Visualizing Class-Agnostic Features of CNNs [92.39082696657874]
We propose an approach to visually interpret CNN features given a set of images by creating corresponding images that depict the most informative features of a specific layer.
Our method uses a dual-objective activation and distance loss, without requiring a generator network nor modifications to the original model.
arXiv Detail & Related papers (2021-01-29T07:46:39Z) - Understanding Graph Isomorphism Network for rs-fMRI Functional
Connectivity Analysis [49.05541693243502]
We develop a framework for analyzing fMRI data using the Graph Isomorphism Network (GIN)
One of the important contributions of this paper is the observation that the GIN is a dual representation of convolutional neural network (CNN) in the graph space.
We exploit CNN-based saliency map techniques for the GNN, which we tailor to the proposed GIN with one-hot encoding.
arXiv Detail & Related papers (2020-01-10T23:40:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.