MISSU: 3D Medical Image Segmentation via Self-distilling TransUNet
- URL: http://arxiv.org/abs/2206.00902v1
- Date: Thu, 2 Jun 2022 07:38:53 GMT
- Title: MISSU: 3D Medical Image Segmentation via Self-distilling TransUNet
- Authors: Nan Wang, Shaohui Lin, Xiaoxiao Li, Ke Li, Yunhang Shen, Yue Gao,
Lizhuang Ma
- Abstract summary: We propose to self-distill a Transformer-based UNet for medical image segmentation.
It simultaneously learns global semantic information and local spatial-detailed features.
Our MISSU achieves the best performance over previous state-of-the-art methods.
- Score: 55.16833099336073
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: U-Nets have achieved tremendous success in medical image segmentation.
Nevertheless, it may suffer limitations in global (long-range) contextual
interactions and edge-detail preservation. In contrast, Transformer has an
excellent ability to capture long-range dependencies by leveraging the
self-attention mechanism into the encoder. Although Transformer was born to
model the long-range dependency on the extracted feature maps, it still suffers
from extreme computational and spatial complexities in processing
high-resolution 3D feature maps. This motivates us to design the efficiently
Transformer-based UNet model and study the feasibility of Transformer-based
network architectures for medical image segmentation tasks. To this end, we
propose to self-distill a Transformer-based UNet for medical image
segmentation, which simultaneously learns global semantic information and local
spatial-detailed features. Meanwhile, a local multi-scale fusion block is first
proposed to refine fine-grained details from the skipped connections in the
encoder by the main CNN stem through self-distillation, only computed during
training and removed at inference with minimal overhead. Extensive experiments
on BraTS 2019 and CHAOS datasets show that our MISSU achieves the best
performance over previous state-of-the-art methods. Code and models are
available at \url{https://github.com/wangn123/MISSU.git}
Related papers
- ParaTransCNN: Parallelized TransCNN Encoder for Medical Image
Segmentation [7.955518153976858]
We propose an advanced 2D feature extraction method by combining the convolutional neural network and Transformer architectures.
Our method is shown with better segmentation accuracy, especially on small organs.
arXiv Detail & Related papers (2024-01-27T05:58:36Z) - TransNorm: Transformer Provides a Strong Spatial Normalization Mechanism
for a Deep Segmentation Model [4.320393382724066]
convolutional neural networks (CNNs) have been the prevailing technique in the medical image processing era.
We propose Trans-Norm, a novel deep segmentation framework which consolidates a Transformer module into both encoder and skip-connections of the standard U-Net.
arXiv Detail & Related papers (2022-07-27T09:54:10Z) - Dynamic Linear Transformer for 3D Biomedical Image Segmentation [2.440109381823186]
Transformer-based neural networks have surpassed promising performance on many biomedical image segmentation tasks.
Main challenge for 3D transformer-based segmentation methods is the quadratic complexity introduced by the self-attention mechanism.
We propose a novel transformer architecture for 3D medical image segmentation using an encoder-decoder style architecture with linear complexity.
arXiv Detail & Related papers (2022-06-01T21:15:01Z) - Swin-Unet: Unet-like Pure Transformer for Medical Image Segmentation [63.46694853953092]
Swin-Unet is an Unet-like pure Transformer for medical image segmentation.
tokenized image patches are fed into the Transformer-based U-shaped decoder-Decoder architecture.
arXiv Detail & Related papers (2021-05-12T09:30:26Z) - Transformers Solve the Limited Receptive Field for Monocular Depth
Prediction [82.90445525977904]
We propose TransDepth, an architecture which benefits from both convolutional neural networks and transformers.
This is the first paper which applies transformers into pixel-wise prediction problems involving continuous labels.
arXiv Detail & Related papers (2021-03-22T18:00:13Z) - CoTr: Efficiently Bridging CNN and Transformer for 3D Medical Image
Segmentation [95.51455777713092]
Convolutional neural networks (CNNs) have been the de facto standard for nowadays 3D medical image segmentation.
We propose a novel framework that efficiently bridges a bf Convolutional neural network and a bf Transformer bf (CoTr) for accurate 3D medical image segmentation.
arXiv Detail & Related papers (2021-03-04T13:34:22Z) - Medical Transformer: Gated Axial-Attention for Medical Image
Segmentation [73.98974074534497]
We study the feasibility of using Transformer-based network architectures for medical image segmentation tasks.
We propose a Gated Axial-Attention model which extends the existing architectures by introducing an additional control mechanism in the self-attention module.
To train the model effectively on medical images, we propose a Local-Global training strategy (LoGo) which further improves the performance.
arXiv Detail & Related papers (2021-02-21T18:35:14Z) - TransUNet: Transformers Make Strong Encoders for Medical Image
Segmentation [78.01570371790669]
Medical image segmentation is an essential prerequisite for developing healthcare systems.
On various medical image segmentation tasks, the u-shaped architecture, also known as U-Net, has become the de-facto standard.
We propose TransUNet, which merits both Transformers and U-Net, as a strong alternative for medical image segmentation.
arXiv Detail & Related papers (2021-02-08T16:10:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.