Swin Deformable Attention U-Net Transformer (SDAUT) for Explainable Fast
MRI
- URL: http://arxiv.org/abs/2207.02390v1
- Date: Tue, 5 Jul 2022 15:56:46 GMT
- Title: Swin Deformable Attention U-Net Transformer (SDAUT) for Explainable Fast
MRI
- Authors: Jiahao Huang, Xiaodan Xing, Zhifan Gao, Guang Yang
- Abstract summary: We propose a new Transformer architecture for solving fast MRI.
We incorporate deformable attention to construe the explainability of our reconstruction model.
Our method has fewer network parameters while revealing explainability.
- Score: 3.2621521013133385
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Fast MRI aims to reconstruct a high fidelity image from partially observed
measurements. Exuberant development in fast MRI using deep learning has been
witnessed recently. Meanwhile, novel deep learning paradigms, e.g., Transformer
based models, are fast-growing in natural language processing and promptly
developed for computer vision and medical image analysis due to their prominent
performance. Nevertheless, due to the complexity of the Transformer, the
application of fast MRI may not be straightforward. The main obstacle is the
computational cost of the self-attention layer, which is the core part of the
Transformer, can be expensive for high resolution MRI inputs. In this study, we
propose a new Transformer architecture for solving fast MRI that coupled
Shifted Windows Transformer with U-Net to reduce the network complexity. We
incorporate deformable attention to construe the explainability of our
reconstruction model. We empirically demonstrate that our method achieves
consistently superior performance on the fast MRI task. Besides, compared to
state-of-the-art Transformer models, our method has fewer network parameters
while revealing explainability. The code is publicly available at
https://github.com/ayanglab/SDAUT.
Related papers
- MindFormer: A Transformer Architecture for Multi-Subject Brain Decoding via fMRI [50.55024115943266]
We introduce a new Transformer architecture called MindFormer to generate fMRI-conditioned feature vectors.
MindFormer incorporates two key innovations: 1) a novel training strategy based on the IP-Adapter to extract semantically meaningful features from fMRI signals, and 2) a subject specific token and linear layer that effectively capture individual differences in fMRI signals.
arXiv Detail & Related papers (2024-05-28T00:36:25Z) - Learning Dynamic MRI Reconstruction with Convolutional Network Assisted
Reconstruction Swin Transformer [0.7802769338493889]
We propose a novel architecture named Reconstruction Swin Transformer (RST) for 4D MRI.
RST inherits the backbone design of the Video Swin Transformer with a novel reconstruction head introduced to restore pixel-wise intensity.
Experimental results in the cardiac 4D MR dataset further substantiate the superiority of RST.
arXiv Detail & Related papers (2023-09-19T00:42:45Z) - Cross-Attention of Disentangled Modalities for 3D Human Mesh Recovery
with Transformers [17.22112222736234]
Transformer encoder architectures have recently achieved state-of-the-art results on monocular 3D human mesh reconstruction.
Due to the large memory overhead and slow inference speed, it is difficult to deploy such models for practical use.
We propose a novel transformer encoder-decoder architecture for 3D human mesh reconstruction from a single image, called FastMETRO.
arXiv Detail & Related papers (2022-07-27T22:54:09Z) - Cross-receptive Focused Inference Network for Lightweight Image
Super-Resolution [64.25751738088015]
Transformer-based methods have shown impressive performance in single image super-resolution (SISR) tasks.
Transformers that need to incorporate contextual information to extract features dynamically are neglected.
We propose a lightweight Cross-receptive Focused Inference Network (CFIN) that consists of a cascade of CT Blocks mixed with CNN and Transformer.
arXiv Detail & Related papers (2022-07-06T16:32:29Z) - MISSU: 3D Medical Image Segmentation via Self-distilling TransUNet [55.16833099336073]
We propose to self-distill a Transformer-based UNet for medical image segmentation.
It simultaneously learns global semantic information and local spatial-detailed features.
Our MISSU achieves the best performance over previous state-of-the-art methods.
arXiv Detail & Related papers (2022-06-02T07:38:53Z) - Cross-Modality High-Frequency Transformer for MR Image Super-Resolution [100.50972513285598]
We build an early effort to build a Transformer-based MR image super-resolution framework.
We consider two-fold domain priors including the high-frequency structure prior and the inter-modality context prior.
We establish a novel Transformer architecture, called Cross-modality high-frequency Transformer (Cohf-T), to introduce such priors into super-resolving the low-resolution images.
arXiv Detail & Related papers (2022-03-29T07:56:55Z) - HUMUS-Net: Hybrid unrolled multi-scale network architecture for
accelerated MRI reconstruction [38.0542877099235]
HUMUS-Net is a hybrid architecture that combines the beneficial implicit bias and efficiency of convolutions with the power of Transformer blocks in an unrolled and multi-scale network.
Our network establishes new state of the art on the largest publicly available MRI dataset, the fastMRI dataset.
arXiv Detail & Related papers (2022-03-15T19:26:29Z) - Fast MRI Reconstruction: How Powerful Transformers Are? [1.523157765626545]
Methods by k-space undersampling and deep learning based reconstruction have been popularised to accelerate the scanning process.
In particular, a generative adversarial network (GAN) based Swin transformer (ST-GAN) was introduced for the fast MRI reconstruction.
We show that transformers work well for the MRI reconstruction from different undersampling conditions.
arXiv Detail & Related papers (2022-01-23T23:41:48Z) - UTNet: A Hybrid Transformer Architecture for Medical Image Segmentation [6.646135062704341]
Transformer architecture has emerged to be successful in a number of natural language processing tasks.
We present UTNet, a powerful hybrid Transformer architecture that integrates self-attention into a convolutional neural network for enhancing medical image segmentation.
arXiv Detail & Related papers (2021-07-02T00:56:27Z) - Adaptive Gradient Balancing for UndersampledMRI Reconstruction and
Image-to-Image Translation [60.663499381212425]
We enhance the image quality by using a Wasserstein Generative Adversarial Network combined with a novel Adaptive Gradient Balancing technique.
In MRI, our method minimizes artifacts, while maintaining a high-quality reconstruction that produces sharper images than other techniques.
arXiv Detail & Related papers (2021-04-05T13:05:22Z) - Medical Transformer: Gated Axial-Attention for Medical Image
Segmentation [73.98974074534497]
We study the feasibility of using Transformer-based network architectures for medical image segmentation tasks.
We propose a Gated Axial-Attention model which extends the existing architectures by introducing an additional control mechanism in the self-attention module.
To train the model effectively on medical images, we propose a Local-Global training strategy (LoGo) which further improves the performance.
arXiv Detail & Related papers (2021-02-21T18:35:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.