DoDNet: Learning to segment multi-organ and tumors from multiple
partially labeled datasets
- URL: http://arxiv.org/abs/2011.10217v1
- Date: Fri, 20 Nov 2020 04:56:39 GMT
- Title: DoDNet: Learning to segment multi-organ and tumors from multiple
partially labeled datasets
- Authors: Jianpeng Zhang, Yutong Xie, Yong Xia, Chunhua Shen
- Abstract summary: We propose a dynamic on-demand network (DoDNet) that learns to segment multiple organs and tumors on partially labelled datasets.
DoDNet consists of a shared encoder-decoder architecture, a task encoding module, a controller for generating dynamic convolution filters, and a single but dynamic segmentation head.
- Score: 102.55303521877933
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Due to the intensive cost of labor and expertise in annotating 3D medical
images at a voxel level, most benchmark datasets are equipped with the
annotations of only one type of organs and/or tumors, resulting in the
so-called partially labeling issue. To address this, we propose a dynamic
on-demand network (DoDNet) that learns to segment multiple organs and tumors on
partially labeled datasets. DoDNet consists of a shared encoder-decoder
architecture, a task encoding module, a controller for generating dynamic
convolution filters, and a single but dynamic segmentation head. The
information of the current segmentation task is encoded as a task-aware prior
to tell the model what the task is expected to solve. Different from existing
approaches which fix kernels after training, the kernels in dynamic head are
generated adaptively by the controller, conditioned on both input image and
assigned task. Thus, DoDNet is able to segment multiple organs and tumors, as
done by multiple networks or a multi-head network, in a much efficient and
flexible manner. We have created a large-scale partially labeled dataset,
termed MOTS, and demonstrated the superior performance of our DoDNet over other
competitors on seven organ and tumor segmentation tasks. We also transferred
the weights pre-trained on MOTS to a downstream multi-organ segmentation task
and achieved state-of-the-art performance. This study provides a general 3D
medical image segmentation model that has been pre-trained on a large-scale
partially labelled dataset and can be extended (after fine-tuning) to
downstream volumetric medical data segmentation tasks. The dataset and code
areavailableat: https://git.io/DoDNet
Related papers
- ESDMR-Net: A Lightweight Network With Expand-Squeeze and Dual Multiscale
Residual Connections for Medical Image Segmentation [7.921517156237902]
This paper presents an expand-squeeze dual multiscale residual network ( ESDMR-Net)
It is a fully convolutional network that is well-suited for resource-constrained computing hardware such as mobile devices.
We present experiments on seven datasets from five distinct examples of applications.
arXiv Detail & Related papers (2023-12-17T02:15:49Z) - Learning from partially labeled data for multi-organ and tumor
segmentation [102.55303521877933]
We propose a Transformer based dynamic on-demand network (TransDoDNet) that learns to segment organs and tumors on multiple datasets.
A dynamic head enables the network to accomplish multiple segmentation tasks flexibly.
We create a large-scale partially labeled Multi-Organ and Tumor benchmark, termed MOTS, and demonstrate the superior performance of our TransDoDNet over other competitors.
arXiv Detail & Related papers (2022-11-13T13:03:09Z) - Dual Multi-scale Mean Teacher Network for Semi-supervised Infection
Segmentation in Chest CT Volume for COVID-19 [76.51091445670596]
Automated detecting lung infections from computed tomography (CT) data plays an important role for combating COVID-19.
Most current COVID-19 infection segmentation methods mainly relied on 2D CT images, which lack 3D sequential constraint.
Existing 3D CT segmentation methods focus on single-scale representations, which do not achieve the multiple level receptive field sizes on 3D volume.
arXiv Detail & Related papers (2022-11-10T13:11:21Z) - Two-Stream Graph Convolutional Network for Intra-oral Scanner Image
Segmentation [133.02190910009384]
We propose a two-stream graph convolutional network (i.e., TSGCN) to handle inter-view confusion between different raw attributes.
Our TSGCN significantly outperforms state-of-the-art methods in 3D tooth (surface) segmentation.
arXiv Detail & Related papers (2022-04-19T10:41:09Z) - UNetFormer: A Unified Vision Transformer Model and Pre-Training
Framework for 3D Medical Image Segmentation [14.873473285148853]
We introduce a unified framework consisting of two architectures, dubbed UNetFormer, with a 3D Swin Transformer-based encoder and Conal Neural Network (CNN) and transformer-based decoders.
In the proposed model, the encoder is linked to the decoder via skip connections at five different resolutions with deep supervision.
We present a methodology for self-supervised pre-training of the encoder backbone via learning to predict randomly masked tokens.
arXiv Detail & Related papers (2022-04-01T17:38:39Z) - Generalized Organ Segmentation by Imitating One-shot Reasoning using
Anatomical Correlation [55.1248480381153]
We propose OrganNet which learns a generalized organ concept from a set of annotated organ classes and then transfer this concept to unseen classes.
We show that OrganNet can effectively resist the wide variations in organ morphology and produce state-of-the-art results in one-shot segmentation task.
arXiv Detail & Related papers (2021-03-30T13:41:12Z) - 3D Segmentation Networks for Excessive Numbers of Classes: Distinct Bone
Segmentation in Upper Bodies [1.2023648183416153]
This paper discusses the intricacies of training a 3D segmentation network in a many-label setting.
We show necessary modifications in network architecture, loss function, and data augmentation.
As a result, we demonstrate the robustness of our method by automatically segmenting over one hundred distinct bones simultaneously in an end-to-end learnt fashion from a CT-scan.
arXiv Detail & Related papers (2020-10-14T12:54:15Z) - Boundary-aware Context Neural Network for Medical Image Segmentation [15.585851505721433]
Medical image segmentation can provide reliable basis for further clinical analysis and disease diagnosis.
Most existing CNNs-based methods produce unsatisfactory segmentation mask without accurate object boundaries.
In this paper, we formulate a boundary-aware context neural network (BA-Net) for 2D medical image segmentation.
arXiv Detail & Related papers (2020-05-03T02:35:49Z) - 3D medical image segmentation with labeled and unlabeled data using
autoencoders at the example of liver segmentation in CT images [58.720142291102135]
This work investigates the potential of autoencoder-extracted features to improve segmentation with a convolutional neural network.
A convolutional autoencoder was used to extract features from unlabeled data and a multi-scale, fully convolutional CNN was used to perform the target task of 3D liver segmentation in CT images.
arXiv Detail & Related papers (2020-03-17T20:20:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.