Learning from partially labeled data for multi-organ and tumor
segmentation
- URL: http://arxiv.org/abs/2211.06894v1
- Date: Sun, 13 Nov 2022 13:03:09 GMT
- Title: Learning from partially labeled data for multi-organ and tumor
segmentation
- Authors: Yutong Xie, Jianpeng Zhang, Yong Xia, Chunhua Shen
- Abstract summary: We propose a Transformer based dynamic on-demand network (TransDoDNet) that learns to segment organs and tumors on multiple datasets.
A dynamic head enables the network to accomplish multiple segmentation tasks flexibly.
We create a large-scale partially labeled Multi-Organ and Tumor benchmark, termed MOTS, and demonstrate the superior performance of our TransDoDNet over other competitors.
- Score: 102.55303521877933
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Medical image benchmarks for the segmentation of organs and tumors suffer
from the partially labeling issue due to its intensive cost of labor and
expertise. Current mainstream approaches follow the practice of one network
solving one task. With this pipeline, not only the performance is limited by
the typically small dataset of a single task, but also the computation cost
linearly increases with the number of tasks. To address this, we propose a
Transformer based dynamic on-demand network (TransDoDNet) that learns to
segment organs and tumors on multiple partially labeled datasets. Specifically,
TransDoDNet has a hybrid backbone that is composed of the convolutional neural
network and Transformer. A dynamic head enables the network to accomplish
multiple segmentation tasks flexibly. Unlike existing approaches that fix
kernels after training, the kernels in the dynamic head are generated
adaptively by the Transformer, which employs the self-attention mechanism to
model long-range organ-wise dependencies and decodes the organ embedding that
can represent each organ. We create a large-scale partially labeled Multi-Organ
and Tumor Segmentation benchmark, termed MOTS, and demonstrate the superior
performance of our TransDoDNet over other competitors on seven organ and tumor
segmentation tasks. This study also provides a general 3D medical image
segmentation model, which has been pre-trained on the large-scale MOTS
benchmark and has demonstrated advanced performance over BYOL, the current
predominant self-supervised learning method. Code will be available at
\url{https://git.io/DoDNet}.
Related papers
- Investigation of Network Architecture for Multimodal Head-and-Neck Tumor
Segmentation [9.441769048218955]
In this study, we analyze, two recently published Transformer-based network architectures for the task of multimodal head-and-tumor segmentation.
Our results showed that modeling long-range dependencies may be helpful in cases where large structures are present and/or large field of view is needed.
For small structures such as head-and-neck tumor, the convolution-based U-Net architecture seemed to perform well, especially when training dataset is small and computational resource is limited.
arXiv Detail & Related papers (2022-12-21T02:35:46Z) - A Transformer-based Generative Adversarial Network for Brain Tumor
Segmentation [4.394247741333439]
We propose a transformer-based generative adversarial network to automatically segment brain tumors with multi-modalities MRI.
Our architecture consists of a generator and a discriminator, which are trained in min-max game progress.
The discriminator we designed is a CNN-based network with multi-scale $L_1$ loss, which is proved to be effective for medical semantic image segmentation.
arXiv Detail & Related papers (2022-07-28T14:55:18Z) - MISSU: 3D Medical Image Segmentation via Self-distilling TransUNet [55.16833099336073]
We propose to self-distill a Transformer-based UNet for medical image segmentation.
It simultaneously learns global semantic information and local spatial-detailed features.
Our MISSU achieves the best performance over previous state-of-the-art methods.
arXiv Detail & Related papers (2022-06-02T07:38:53Z) - Swin UNETR: Swin Transformers for Semantic Segmentation of Brain Tumors
in MRI Images [7.334185314342017]
We propose a novel segmentation model termed Swin UNEt TRansformers (Swin UNETR)
The model extracts features at five different resolutions by utilizing shifted windows for computing self-attention.
We have participated in BraTS 2021 segmentation challenge, and our proposed model ranks among the top-performing approaches in the validation phase.
arXiv Detail & Related papers (2022-01-04T18:01:34Z) - Cross-Modality Brain Tumor Segmentation via Bidirectional
Global-to-Local Unsupervised Domain Adaptation [61.01704175938995]
In this paper, we propose a novel Bidirectional Global-to-Local (BiGL) adaptation framework under a UDA scheme.
Specifically, a bidirectional image synthesis and segmentation module is proposed to segment the brain tumor.
The proposed method outperforms several state-of-the-art unsupervised domain adaptation methods by a large margin.
arXiv Detail & Related papers (2021-05-17T10:11:45Z) - Generalized Organ Segmentation by Imitating One-shot Reasoning using
Anatomical Correlation [55.1248480381153]
We propose OrganNet which learns a generalized organ concept from a set of annotated organ classes and then transfer this concept to unseen classes.
We show that OrganNet can effectively resist the wide variations in organ morphology and produce state-of-the-art results in one-shot segmentation task.
arXiv Detail & Related papers (2021-03-30T13:41:12Z) - TransUNet: Transformers Make Strong Encoders for Medical Image
Segmentation [78.01570371790669]
Medical image segmentation is an essential prerequisite for developing healthcare systems.
On various medical image segmentation tasks, the u-shaped architecture, also known as U-Net, has become the de-facto standard.
We propose TransUNet, which merits both Transformers and U-Net, as a strong alternative for medical image segmentation.
arXiv Detail & Related papers (2021-02-08T16:10:50Z) - DoDNet: Learning to segment multi-organ and tumors from multiple
partially labeled datasets [102.55303521877933]
We propose a dynamic on-demand network (DoDNet) that learns to segment multiple organs and tumors on partially labelled datasets.
DoDNet consists of a shared encoder-decoder architecture, a task encoding module, a controller for generating dynamic convolution filters, and a single but dynamic segmentation head.
arXiv Detail & Related papers (2020-11-20T04:56:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.