MetricUNet: Synergistic Image- and Voxel-Level Learning for Precise CT
Prostate Segmentation via Online Sampling
- URL: http://arxiv.org/abs/2005.07462v4
- Date: Sat, 23 Jan 2021 17:18:35 GMT
- Title: MetricUNet: Synergistic Image- and Voxel-Level Learning for Precise CT
Prostate Segmentation via Online Sampling
- Authors: Kelei He, Chunfeng Lian, Ehsan Adeli, Jing Huo, Yang Gao, Bing Zhang,
Junfeng Zhang, Dinggang Shen
- Abstract summary: We propose a two-stage framework, with the first stage to quickly localize the prostate region and the second stage to precisely segment the prostate.
We introduce a novel online metric learning module through voxel-wise sampling in the multi-task network.
Our method can effectively learn more representative voxel-level features compared with the conventional learning methods with cross-entropy or Dice loss.
- Score: 66.01558025094333
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Fully convolutional networks (FCNs), including UNet and VNet, are widely-used
network architectures for semantic segmentation in recent studies. However,
conventional FCN is typically trained by the cross-entropy or Dice loss, which
only calculates the error between predictions and ground-truth labels for
pixels individually. This often results in non-smooth neighborhoods in the
predicted segmentation. To address this problem, we propose a two-stage
framework, with the first stage to quickly localize the prostate region and the
second stage to precisely segment the prostate by a multi-task UNet
architecture. We introduce a novel online metric learning module through
voxel-wise sampling in the multi-task network. Therefore, the proposed network
has a dual-branch architecture that tackles two tasks: 1) a segmentation
sub-network aiming to generate the prostate segmentation, and 2) a voxel-metric
learning sub-network aiming to improve the quality of the learned feature space
supervised by a metric loss. Specifically, the voxel-metric learning
sub-network samples tuples (including triplets and pairs) in voxel-level
through the intermediate feature maps. Unlike conventional deep metric learning
methods that generate triplets or pairs in image-level before the training
phase, our proposed voxel-wise tuples are sampled in an online manner and
operated in an end-to-end fashion via multi-task learning. To evaluate the
proposed method, we implement extensive experiments on a real CT image dataset
consisting of 339 patients. The ablation studies show that our method can
effectively learn more representative voxel-level features compared with the
conventional learning methods with cross-entropy or Dice loss. And the
comparisons show that the proposed method outperforms the state-of-the-art
methods by a reasonable margin.
Related papers
- IDEAL: Improved DEnse locAL Contrastive Learning for Semi-Supervised
Medical Image Segmentation [3.6748639131154315]
We extend the concept of metric learning to the segmentation task.
We propose a simple convolutional projection head for obtaining dense pixel-level features.
A bidirectional regularization mechanism involving two-stream regularization training is devised for the downstream task.
arXiv Detail & Related papers (2022-10-26T23:11:02Z) - PA-Seg: Learning from Point Annotations for 3D Medical Image
Segmentation using Contextual Regularization and Cross Knowledge Distillation [14.412073730567137]
We propose to annotate a segmentation target with only seven points in 3D medical images, and design a two-stage weakly supervised learning framework PA-Seg.
In the first stage, we employ geodesic distance transform to expand the seed points to provide more supervision signal.
In the second stage, we use predictions obtained by the model pre-trained in the first stage as pseudo labels.
arXiv Detail & Related papers (2022-08-11T07:00:33Z) - Two-Stream Graph Convolutional Network for Intra-oral Scanner Image
Segmentation [133.02190910009384]
We propose a two-stream graph convolutional network (i.e., TSGCN) to handle inter-view confusion between different raw attributes.
Our TSGCN significantly outperforms state-of-the-art methods in 3D tooth (surface) segmentation.
arXiv Detail & Related papers (2022-04-19T10:41:09Z) - Multi-scale and Cross-scale Contrastive Learning for Semantic
Segmentation [5.281694565226513]
We apply contrastive learning to enhance the discriminative power of the multi-scale features extracted by semantic segmentation networks.
By first mapping the encoder's multi-scale representations to a common feature space, we instantiate a novel form of supervised local-global constraint.
arXiv Detail & Related papers (2022-03-25T01:24:24Z) - TSGCNet: Discriminative Geometric Feature Learning with Two-Stream
GraphConvolutional Network for 3D Dental Model Segmentation [141.2690520327948]
We propose a two-stream graph convolutional network (TSGCNet) to learn multi-view information from different geometric attributes.
We evaluate our proposed TSGCNet on a real-patient dataset of dental models acquired by 3D intraoral scanners.
arXiv Detail & Related papers (2020-12-26T08:02:56Z) - Fast accuracy estimation of deep learning based multi-class musical
source separation [79.10962538141445]
We propose a method to evaluate the separability of instruments in any dataset without training and tuning a neural network.
Based on the oracle principle with an ideal ratio mask, our approach is an excellent proxy to estimate the separation performances of state-of-the-art deep learning approaches.
arXiv Detail & Related papers (2020-10-19T13:05:08Z) - A Transductive Multi-Head Model for Cross-Domain Few-Shot Learning [72.30054522048553]
We present a new method, Transductive Multi-Head Few-Shot learning (TMHFS), to address the Cross-Domain Few-Shot Learning challenge.
The proposed methods greatly outperform the strong baseline, fine-tuning, on four different target domains.
arXiv Detail & Related papers (2020-06-08T02:39:59Z) - Unpaired Multi-modal Segmentation via Knowledge Distillation [77.39798870702174]
We propose a novel learning scheme for unpaired cross-modality image segmentation.
In our method, we heavily reuse network parameters, by sharing all convolutional kernels across CT and MRI.
We have extensively validated our approach on two multi-class segmentation problems.
arXiv Detail & Related papers (2020-01-06T20:03:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.