SAR: Scale-Aware Restoration Learning for 3D Tumor Segmentation
- URL: http://arxiv.org/abs/2010.06107v2
- Date: Thu, 5 Aug 2021 09:25:04 GMT
- Title: SAR: Scale-Aware Restoration Learning for 3D Tumor Segmentation
- Authors: Xiaoman Zhang, Shixiang Feng, Yuhang Zhou, Ya Zhang and Yanfeng Wang
- Abstract summary: We propose Scale-Aware Restoration (SAR) for 3D tumor segmentation.
A novel proxy task, i.e. scale discrimination, is formulated to pre-train the 3D neural network combined with the self-restoration task.
We demonstrate the effectiveness of our methods on two downstream tasks: i.e. Brain tumor segmentation, ii. Pancreas tumor segmentation.
- Score: 23.384259038420005
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Automatic and accurate tumor segmentation on medical images is in high demand
to assist physicians with diagnosis and treatment. However, it is difficult to
obtain massive amounts of annotated training data required by the deep-learning
models as the manual delineation process is often tedious and expertise
required. Although self-supervised learning (SSL) scheme has been widely
adopted to address this problem, most SSL methods focus only on global
structure information, ignoring the key distinguishing features of tumor
regions: local intensity variation and large size distribution. In this paper,
we propose Scale-Aware Restoration (SAR), a SSL method for 3D tumor
segmentation. Specifically, a novel proxy task, i.e. scale discrimination, is
formulated to pre-train the 3D neural network combined with the
self-restoration task. Thus, the pre-trained model learns multi-level local
representations through multi-scale inputs. Moreover, an adversarial learning
module is further introduced to learn modality invariant representations from
multiple unlabeled source datasets. We demonstrate the effectiveness of our
methods on two downstream tasks: i) Brain tumor segmentation, ii) Pancreas
tumor segmentation. Compared with the state-of-the-art 3D SSL methods, our
proposed approach can significantly improve the segmentation accuracy. Besides,
we analyze its advantages from multiple perspectives such as data efficiency,
performance, and convergence speed.
Related papers
- MBDRes-U-Net: Multi-Scale Lightweight Brain Tumor Segmentation Network [0.0]
This study proposes the MBDRes-U-Net model using the three-dimensional (3D) U-Net framework, which integrates multibranch residual blocks and fused attention into the model.
The computational burden of the model is reduced by the branch strategy, which effectively uses the rich local features in multimodal images.
arXiv Detail & Related papers (2024-11-04T09:03:43Z) - Improving 3D Medical Image Segmentation at Boundary Regions using Local Self-attention and Global Volume Mixing [14.0825980706386]
Volumetric medical image segmentation is a fundamental problem in medical image analysis where the objective is to accurately classify a given 3D volumetric medical image with voxel-level precision.
In this work, we propose a novel hierarchical encoder-decoder-based framework that strives to explicitly capture the local and global dependencies for 3D medical image segmentation.
The proposed framework exploits local volume-based self-attention to encode the local dependencies at high resolution and introduces a novel volumetric-mixer to capture the global dependencies at low-resolution feature representations.
arXiv Detail & Related papers (2024-10-20T11:08:38Z) - Enhancing Weakly Supervised 3D Medical Image Segmentation through
Probabilistic-aware Learning [52.249748801637196]
3D medical image segmentation is a challenging task with crucial implications for disease diagnosis and treatment planning.
Recent advances in deep learning have significantly enhanced fully supervised medical image segmentation.
We propose a novel probabilistic-aware weakly supervised learning pipeline, specifically designed for 3D medical imaging.
arXiv Detail & Related papers (2024-03-05T00:46:53Z) - Self-supervised learning via inter-modal reconstruction and feature
projection networks for label-efficient 3D-to-2D segmentation [4.5206601127476445]
We propose a novel convolutional neural network (CNN) and self-supervised learning (SSL) method for label-efficient 3D-to-2D segmentation.
Results on different datasets demonstrate that the proposed CNN significantly improves the state of the art in scenarios with limited labeled data by up to 8% in Dice score.
arXiv Detail & Related papers (2023-07-06T14:16:25Z) - S3M: Scalable Statistical Shape Modeling through Unsupervised
Correspondences [91.48841778012782]
We propose an unsupervised method to simultaneously learn local and global shape structures across population anatomies.
Our pipeline significantly improves unsupervised correspondence estimation for SSMs compared to baseline methods.
Our method is robust enough to learn from noisy neural network predictions, potentially enabling scaling SSMs to larger patient populations.
arXiv Detail & Related papers (2023-04-15T09:39:52Z) - PCRLv2: A Unified Visual Information Preservation Framework for
Self-supervised Pre-training in Medical Image Analysis [56.63327669853693]
We propose to incorporate the task of pixel restoration for explicitly encoding more pixel-level information into high-level semantics.
We also address the preservation of scale information, a powerful tool in aiding image understanding.
The proposed unified SSL framework surpasses its self-supervised counterparts on various tasks.
arXiv Detail & Related papers (2023-01-02T17:47:27Z) - A unified 3D framework for Organs at Risk Localization and Segmentation
for Radiation Therapy Planning [56.52933974838905]
Current medical workflow requires manual delineation of organs-at-risk (OAR)
In this work, we aim to introduce a unified 3D pipeline for OAR localization-segmentation.
Our proposed framework fully enables the exploitation of 3D context information inherent in medical imaging.
arXiv Detail & Related papers (2022-03-01T17:08:41Z) - Cross-Modality Deep Feature Learning for Brain Tumor Segmentation [158.8192041981564]
This paper proposes a novel cross-modality deep feature learning framework to segment brain tumors from the multi-modality MRI data.
The core idea is to mine rich patterns across the multi-modality data to make up for the insufficient data scale.
Comprehensive experiments are conducted on the BraTS benchmarks, which show that the proposed cross-modality deep feature learning framework can effectively improve the brain tumor segmentation performance.
arXiv Detail & Related papers (2022-01-07T07:46:01Z) - PGL: Prior-Guided Local Self-supervised Learning for 3D Medical Image
Segmentation [87.50205728818601]
We propose a PriorGuided Local (PGL) self-supervised model that learns the region-wise local consistency in the latent feature space.
Our PGL model learns the distinctive representations of local regions, and hence is able to retain structural information.
arXiv Detail & Related papers (2020-11-25T11:03:11Z) - 3D Segmentation Networks for Excessive Numbers of Classes: Distinct Bone
Segmentation in Upper Bodies [1.2023648183416153]
This paper discusses the intricacies of training a 3D segmentation network in a many-label setting.
We show necessary modifications in network architecture, loss function, and data augmentation.
As a result, we demonstrate the robustness of our method by automatically segmenting over one hundred distinct bones simultaneously in an end-to-end learnt fashion from a CT-scan.
arXiv Detail & Related papers (2020-10-14T12:54:15Z) - Embedding Task Knowledge into 3D Neural Networks via Self-supervised
Learning [21.902313057142905]
Self-supervised learning (SSL) is a potential solution for deficient annotated data.
We propose a novel SSL approach for 3D medical image classification, namely Task-related Contrastive Prediction Coding ( TCPC)
TCPC embeds task knowledge into training 3D neural networks.
arXiv Detail & Related papers (2020-06-10T12:37:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.