Instance Segmentation of Unlabeled Modalities via Cyclic Segmentation
GAN
- URL: http://arxiv.org/abs/2204.03082v1
- Date: Wed, 6 Apr 2022 20:46:39 GMT
- Title: Instance Segmentation of Unlabeled Modalities via Cyclic Segmentation
GAN
- Authors: Leander Lauenburg, Zudi Lin, Ruihan Zhang, M\'arcia dos Santos, Siyu
Huang, Ignacio Arganda-Carreras, Edward S. Boyden, Hanspeter Pfister, Donglai
Wei
- Abstract summary: We propose a novel Cyclic Generative Adrial Network (CySGAN) that conducts image translation and instance segmentation jointly.
We benchmark our approach on the task of 3D neuronal nuclei segmentation with annotated electron microscopy (EM) images and unlabeled expansion microscopy (ExM) data.
- Score: 27.936725483892076
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Instance segmentation for unlabeled imaging modalities is a challenging but
essential task as collecting expert annotation can be expensive and
time-consuming. Existing works segment a new modality by either deploying a
pre-trained model optimized on diverse training data or conducting domain
translation and image segmentation as two independent steps. In this work, we
propose a novel Cyclic Segmentation Generative Adversarial Network (CySGAN)
that conducts image translation and instance segmentation jointly using a
unified framework. Besides the CycleGAN losses for image translation and
supervised losses for the annotated source domain, we introduce additional
self-supervised and segmentation-based adversarial objectives to improve the
model performance by leveraging unlabeled target domain images. We benchmark
our approach on the task of 3D neuronal nuclei segmentation with annotated
electron microscopy (EM) images and unlabeled expansion microscopy (ExM) data.
Our CySGAN outperforms both pretrained generalist models and the baselines that
sequentially conduct image translation and segmentation. Our implementation and
the newly collected, densely annotated ExM nuclei dataset, named NucExM, are
available at https://connectomics-bazaar.github.io/proj/CySGAN/index.html.
Related papers
- Unsupervised Universal Image Segmentation [59.0383635597103]
We propose an Unsupervised Universal model (U2Seg) adept at performing various image segmentation tasks.
U2Seg generates pseudo semantic labels for these segmentation tasks via leveraging self-supervised models.
We then self-train the model on these pseudo semantic labels, yielding substantial performance gains.
arXiv Detail & Related papers (2023-12-28T18:59:04Z) - SEGIC: Unleashing the Emergent Correspondence for In-Context Segmentation [87.18373801829314]
In-context segmentation aims at segmenting novel images using a few labeled example images, termed as "in-context examples"
We propose SEGIC, an end-to-end segment-in-context framework built upon a single vision foundation model (VFM)
SEGIC is a straightforward yet effective approach that yields state-of-the-art performance on one-shot segmentation benchmarks.
arXiv Detail & Related papers (2023-11-24T18:59:42Z) - SamDSK: Combining Segment Anything Model with Domain-Specific Knowledge
for Semi-Supervised Learning in Medical Image Segmentation [27.044797468878837]
The Segment Anything Model (SAM) exhibits a capability to segment a wide array of objects in natural images.
We propose a novel method that combines the SAM with domain-specific knowledge for reliable utilization of unlabeled images.
Our work initiates a new direction of semi-supervised learning for medical image segmentation.
arXiv Detail & Related papers (2023-08-26T04:46:10Z) - Ensembling Instance and Semantic Segmentation for Panoptic Segmentation [0.0]
Methods first performs instance segmentation and semantic segmentation separately, then combines the two to generate panoptic segmentation results.
We add several expert models of Mask R-CNN in instance segmentation to tackle the data imbalance problem in the training data.
In semantic segmentation, we trained several models with various backbones and use an ensemble strategy which further boosts the segmentation results.
arXiv Detail & Related papers (2023-04-20T14:02:01Z) - Self-Supervised Correction Learning for Semi-Supervised Biomedical Image
Segmentation [84.58210297703714]
We propose a self-supervised correction learning paradigm for semi-supervised biomedical image segmentation.
We design a dual-task network, including a shared encoder and two independent decoders for segmentation and lesion region inpainting.
Experiments on three medical image segmentation datasets for different tasks demonstrate the outstanding performance of our method.
arXiv Detail & Related papers (2023-01-12T08:19:46Z) - HYLDA: End-to-end Hybrid Learning Domain Adaptation for LiDAR Semantic
Segmentation [13.87939140266266]
This paper addresses the problem of training a LiDAR semantic segmentation network using a fully-labeled source dataset and a target dataset that only has a small number of labels.
We develop a novel image-to-image translation engine, and couple it with a LiDAR semantic segmentation network, resulting in an integrated domain adaptation architecture we call HYLDA.
arXiv Detail & Related papers (2022-01-14T18:13:09Z) - Boosting Few-shot Semantic Segmentation with Transformers [81.43459055197435]
TRansformer-based Few-shot Semantic segmentation method (TRFS)
Our model consists of two modules: Global Enhancement Module (GEM) and Local Enhancement Module (LEM)
arXiv Detail & Related papers (2021-08-04T20:09:21Z) - ICMSC: Intra- and Cross-modality Semantic Consistency for Unsupervised
Domain Adaptation on Hip Joint Bone Segmentation [1.4148874598036136]
We propose intra- and cross-modality semantic consistency (ICMSC) for UDA.
Our proposed method achieves an average DICE of 81.61% on the acetabulum and 88.16% on the proximal femur.
Without UDA, a model trained on CT for hip joint bone segmentation is non-transferable to MRI and has almost zero-DICE segmentation.
arXiv Detail & Related papers (2020-12-23T09:58:38Z) - Group-Wise Semantic Mining for Weakly Supervised Semantic Segmentation [49.90178055521207]
This work addresses weakly supervised semantic segmentation (WSSS), with the goal of bridging the gap between image-level annotations and pixel-level segmentation.
We formulate WSSS as a novel group-wise learning task that explicitly models semantic dependencies in a group of images to estimate more reliable pseudo ground-truths.
In particular, we devise a graph neural network (GNN) for group-wise semantic mining, wherein input images are represented as graph nodes.
arXiv Detail & Related papers (2020-12-09T12:40:13Z) - Pairwise Relation Learning for Semi-supervised Gland Segmentation [90.45303394358493]
We propose a pairwise relation-based semi-supervised (PRS2) model for gland segmentation on histology images.
This model consists of a segmentation network (S-Net) and a pairwise relation network (PR-Net)
We evaluate our model against five recent methods on the GlaS dataset and three recent methods on the CRAG dataset.
arXiv Detail & Related papers (2020-08-06T15:02:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.