Robust One-shot Segmentation of Brain Tissues via Image-aligned Style
Transformation
- URL: http://arxiv.org/abs/2211.14521v3
- Date: Wed, 30 Nov 2022 15:05:06 GMT
- Title: Robust One-shot Segmentation of Brain Tissues via Image-aligned Style
Transformation
- Authors: Jinxin Lv, Xiaoyu Zeng, Sheng Wang, Ran Duan, Zhiwei Wang, and Qiang
Li
- Abstract summary: We propose a novel image-aligned style transformation to reinforce the dual-model iterative learning for one-shot segmentation of brain tissues.
Experimental results on two public datasets demonstrate 1) a competitive segmentation performance of our method compared to the fully-supervised method, and 2) a superior performance over other state-of-the-art with an increase of average Dice by up to 4.67%.
- Score: 13.430851964063534
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: One-shot segmentation of brain tissues is typically a dual-model iterative
learning: a registration model (reg-model) warps a carefully-labeled atlas onto
unlabeled images to initialize their pseudo masks for training a segmentation
model (seg-model); the seg-model revises the pseudo masks to enhance the
reg-model for a better warping in the next iteration. However, there is a key
weakness in such dual-model iteration that the spatial misalignment inevitably
caused by the reg-model could misguide the seg-model, which makes it converge
on an inferior segmentation performance eventually. In this paper, we propose a
novel image-aligned style transformation to reinforce the dual-model iterative
learning for robust one-shot segmentation of brain tissues. Specifically, we
first utilize the reg-model to warp the atlas onto an unlabeled image, and then
employ the Fourier-based amplitude exchange with perturbation to transplant the
style of the unlabeled image into the aligned atlas. This allows the subsequent
seg-model to learn on the aligned and style-transferred copies of the atlas
instead of unlabeled images, which naturally guarantees the correct spatial
correspondence of an image-mask training pair, without sacrificing the
diversity of intensity patterns carried by the unlabeled images. Furthermore,
we introduce a feature-aware content consistency in addition to the image-level
similarity to constrain the reg-model for a promising initialization, which
avoids the collapse of image-aligned style transformation in the first
iteration. Experimental results on two public datasets demonstrate 1) a
competitive segmentation performance of our method compared to the
fully-supervised method, and 2) a superior performance over other
state-of-the-art with an increase of average Dice by up to 4.67%. The source
code is available at: https://github.com/JinxLv/One-shot-segmentation-via-IST.
Related papers
- StyleSeg V2: Towards Robust One-shot Segmentation of Brain Tissue via Optimization-free Registration Error Perception [15.118588568951253]
One-shot segmentation of brain tissue requires training registration-segmentation (reg-seg) dual-model iteratively.
Recent StyleSeg bypasses this bottleneck by replacing the unlabeled images with their warped copies of atlas.
Here, we present StyleSeg V2, inherited from StyleSeg but granted the ability of perceiving the registration errors.
arXiv Detail & Related papers (2024-05-06T06:45:23Z) - FreeSeg-Diff: Training-Free Open-Vocabulary Segmentation with Diffusion Models [56.71672127740099]
We focus on the task of image segmentation, which is traditionally solved by training models on closed-vocabulary datasets.
We leverage different and relatively small-sized, open-source foundation models for zero-shot open-vocabulary segmentation.
Our approach (dubbed FreeSeg-Diff), which does not rely on any training, outperforms many training-based approaches on both Pascal VOC and COCO datasets.
arXiv Detail & Related papers (2024-03-29T10:38:25Z) - ZoDi: Zero-Shot Domain Adaptation with Diffusion-Based Image Transfer [13.956618446530559]
This paper proposes a zero-shot domain adaptation method based on diffusion models, called ZoDi.
First, we utilize an off-the-shelf diffusion model to synthesize target-like images by transferring the domain of source images to the target domain.
Secondly, we train the model using both source images and synthesized images with the original representations to learn domain-robust representations.
arXiv Detail & Related papers (2024-03-20T14:58:09Z) - MixReorg: Cross-Modal Mixed Patch Reorganization is a Good Mask Learner
for Open-World Semantic Segmentation [110.09800389100599]
We propose MixReorg, a novel and straightforward pre-training paradigm for semantic segmentation.
Our approach involves generating fine-grained patch-text pairs data by mixing image patches while preserving the correspondence between patches and text.
With MixReorg as a mask learner, conventional text-supervised semantic segmentation models can achieve highly generalizable pixel-semantic alignment ability.
arXiv Detail & Related papers (2023-08-09T09:35:16Z) - CorrMatch: Label Propagation via Correlation Matching for
Semi-Supervised Semantic Segmentation [73.89509052503222]
This paper presents a simple but performant semi-supervised semantic segmentation approach, called CorrMatch.
We observe that the correlation maps not only enable clustering pixels of the same category easily but also contain good shape information.
We propose to conduct pixel propagation by modeling the pairwise similarities of pixels to spread the high-confidence pixels and dig out more.
Then, we perform region propagation to enhance the pseudo labels with accurate class-agnostic masks extracted from the correlation maps.
arXiv Detail & Related papers (2023-06-07T10:02:29Z) - High-fidelity Pseudo-labels for Boosting Weakly-Supervised Segmentation [17.804090651425955]
Image-level weakly-supervised segmentation (WSSS) reduces the usually vast data annotation cost by surrogate segmentation masks during training.
Our work is based on two techniques for improving CAMs; importance sampling, which is a substitute for GAP, and the feature similarity loss.
We reformulate both techniques based on binomial posteriors of multiple independent binary problems.
This has two benefits; their performance is improved and they become more general, resulting in an add-on method that can boost virtually any WSSS method.
arXiv Detail & Related papers (2023-04-05T17:43:57Z) - Cross-modal tumor segmentation using generative blending augmentation and self training [1.6440045168835438]
We propose a cross-modal segmentation method based on conventional image synthesis boosted by a new data augmentation technique.
Generative Blending Augmentation (GBA) learns representative generative features from a single training image to realistically diversify tumor appearances.
The proposed solution ranked first for vestibular schwannoma (VS) segmentation during the validation and test phases of the MICCAI CrossMoDA 2022 challenge.
arXiv Detail & Related papers (2023-04-04T11:01:46Z) - Unified Mask Embedding and Correspondence Learning for Self-Supervised
Video Segmentation [76.40565872257709]
We develop a unified framework which simultaneously models cross-frame dense correspondence for locally discriminative feature learning.
It is able to directly learn to perform mask-guided sequential segmentation from unlabeled videos.
Our algorithm sets state-of-the-arts on two standard benchmarks (i.e., DAVIS17 and YouTube-VOS)
arXiv Detail & Related papers (2023-03-17T16:23:36Z) - Learning Open-vocabulary Semantic Segmentation Models From Natural
Language Supervision [49.905448429974804]
We consider the problem of open-vocabulary semantic segmentation (OVS), which aims to segment objects of arbitrary classes instead of pre-defined, closed-set categories.
We propose a transformer-based model for OVS, termed as OVSegmentor, which exploits web-crawled image-text pairs for pre-training.
Our model achieves superior segmentation results over the state-of-the-art method by using only 3% data (4M vs 134M) for pre-training.
arXiv Detail & Related papers (2023-01-22T13:10:05Z) - Seed the Views: Hierarchical Semantic Alignment for Contrastive
Representation Learning [116.91819311885166]
We propose a hierarchical semantic alignment strategy via expanding the views generated by a single image to textbfCross-samples and Multi-level representation.
Our method, termed as CsMl, has the ability to integrate multi-level visual representations across samples in a robust way.
arXiv Detail & Related papers (2020-12-04T17:26:24Z) - Importance of Self-Consistency in Active Learning for Semantic
Segmentation [31.392212891018655]
We show that self-consistency can be a powerful source of self-supervision to improve the performance of a data-driven model with access to only a small amount of labeled data.
In our proposed active learning framework, we iteratively extract small image patches that need to be labeled.
We are able to find the image patches over which the current model struggles the most to classify.
arXiv Detail & Related papers (2020-08-04T22:18:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.