Training a universal instance segmentation network for live cell images
of various cell types and imaging modalities
- URL: http://arxiv.org/abs/2207.14347v1
- Date: Thu, 28 Jul 2022 18:57:30 GMT
- Title: Training a universal instance segmentation network for live cell images
of various cell types and imaging modalities
- Authors: Tianqi Guo, Yin Wang, Luis Solorio, Jan P. Allebach
- Abstract summary: We present an attempt to train a universal segmentation network for various cell types and imaging modalities.
We modified the traditional binary training targets to include three classes for direct instance segmentation.
Our method was evaluated as the best runner up during the initial submission for the primary track, and also secured the 3rd place in an additional round of competition.
- Score: 10.644558286623813
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We share our recent findings in an attempt to train a universal segmentation
network for various cell types and imaging modalities. Our method was built on
the generalized U-Net architecture, which allows the evaluation of each
component individually. We modified the traditional binary training targets to
include three classes for direct instance segmentation. Detailed experiments
were performed regarding training schemes, training settings, network
backbones, and individual modules on the segmentation performance. Our proposed
training scheme draws minibatches in turn from each dataset, and the gradients
are accumulated before an optimization step. We found that the key to training
a universal network is all-time supervision on all datasets, and it is
necessary to sample each dataset in an unbiased way. Our experiments also
suggest that there might exist common features to define cell boundaries across
cell types and imaging modalities, which could allow application of trained
models to totally unseen datasets. A few training tricks can further boost the
segmentation performance, including uneven class weights in the cross-entropy
loss function, well-designed learning rate scheduler, larger image crops for
contextual information, and additional loss terms for unbalanced classes. We
also found that segmentation performance can benefit from group normalization
layer and Atrous Spatial Pyramid Pooling module, thanks to their more reliable
statistics estimation and improved semantic understanding, respectively. We
participated in the 6th Cell Tracking Challenge (CTC) held at IEEE
International Symposium on Biomedical Imaging (ISBI) 2021 using one of the
developed variants. Our method was evaluated as the best runner up during the
initial submission for the primary track, and also secured the 3rd place in an
additional round of competition in preparation for the summary publication.
Related papers
- Few-Shot Learning for Annotation-Efficient Nucleus Instance Segmentation [50.407071700154674]
We propose to formulate annotation-efficient nucleus instance segmentation from the perspective of few-shot learning (FSL)
Our work was motivated by that, with the prosperity of computational pathology, an increasing number of fully-annotated datasets are publicly accessible.
Extensive experiments on a couple of publicly accessible datasets demonstrate that SGFSIS can outperform other annotation-efficient learning baselines.
arXiv Detail & Related papers (2024-02-26T03:49:18Z) - A-Eval: A Benchmark for Cross-Dataset Evaluation of Abdominal
Multi-Organ Segmentation [38.644744669074775]
We introduce A-Eval, a benchmark for the cross-dataset Evaluation ('Eval') of Abdominal ('A') multi-organ segmentation.
We employ training sets from four large-scale public datasets: FLARE22, AMOS, WORD, and TotalSegmentator.
We evaluate the generalizability of various models using the A-Eval benchmark, with a focus on diverse data usage scenarios.
arXiv Detail & Related papers (2023-09-07T17:59:50Z) - MIS-FM: 3D Medical Image Segmentation using Foundation Models Pretrained
on a Large-Scale Unannotated Dataset [14.823114726604853]
We propose a novel self-supervised learning strategy named Volume Fusion (VF) for pretraining 3D segmentation models.
VF forces the model to predict the fusion coefficient of each voxel, which is formulated as a self-supervised segmentation task without manual annotations.
experiments with different downstream segmentation targets including head and neck organs, thoracic/abdominal organs showed that our pretrained model largely outperformed training from scratch.
arXiv Detail & Related papers (2023-06-29T13:22:13Z) - Joint Optimization of Class-Specific Training- and Test-Time Data
Augmentation in Segmentation [35.41274775082237]
This paper presents an effective and general data augmentation framework for medical image segmentation.
We adopt a computationally efficient and data-efficient gradient-based meta-learning scheme to align the distribution of training and validation data.
We demonstrate the effectiveness of our method on four medical image segmentation tasks with two state-of-the-art segmentation models, DeepMedic and nnU-Net.
arXiv Detail & Related papers (2023-05-30T14:48:45Z) - Towards All-in-one Pre-training via Maximizing Multi-modal Mutual
Information [77.80071279597665]
We propose an all-in-one single-stage pre-training approach, named Maximizing Multi-modal Mutual Information Pre-training (M3I Pre-training)
Our approach achieves better performance than previous pre-training methods on various vision benchmarks, including ImageNet classification, object detection, LVIS long-tailed object detection, and ADE20k semantic segmentation.
arXiv Detail & Related papers (2022-11-17T18:59:49Z) - Unsupervised Domain Adaptation with Histogram-gated Image Translation
for Delayered IC Image Analysis [2.720699926154399]
Histogram-gated Image Translation (HGIT) is an unsupervised domain adaptation framework which transforms images from a given source dataset to the domain of a target dataset.
Our method achieves the best performance compared to the reported domain adaptation techniques, and is also reasonably close to the fully supervised benchmark.
arXiv Detail & Related papers (2022-09-27T15:53:22Z) - Incremental Learning Meets Transfer Learning: Application to Multi-site
Prostate MRI Segmentation [16.50535949349874]
We propose a novel multi-site segmentation framework called incremental-transfer learning (ITL)
ITL learns a model from multi-site datasets in an end-to-end sequential fashion.
We show for the first time that leveraging our ITL training scheme is able to alleviate challenging catastrophic problems in incremental learning.
arXiv Detail & Related papers (2022-06-03T02:32:01Z) - MSeg: A Composite Dataset for Multi-domain Semantic Segmentation [100.17755160696939]
We present MSeg, a composite dataset that unifies semantic segmentation datasets from different domains.
We reconcile the generalization and bring the pixel-level annotations into alignment by relabeling more than 220,000 object masks in more than 80,000 images.
A model trained on MSeg ranks first on the WildDash-v1 leaderboard for robust semantic segmentation, with no exposure to WildDash data during training.
arXiv Detail & Related papers (2021-12-27T16:16:35Z) - Multi-dataset Pretraining: A Unified Model for Semantic Segmentation [97.61605021985062]
We propose a unified framework, termed as Multi-Dataset Pretraining, to take full advantage of the fragmented annotations of different datasets.
This is achieved by first pretraining the network via the proposed pixel-to-prototype contrastive loss over multiple datasets.
In order to better model the relationship among images and classes from different datasets, we extend the pixel level embeddings via cross dataset mixing.
arXiv Detail & Related papers (2021-06-08T06:13:11Z) - Revisiting LSTM Networks for Semi-Supervised Text Classification via
Mixed Objective Function [106.69643619725652]
We develop a training strategy that allows even a simple BiLSTM model, when trained with cross-entropy loss, to achieve competitive results.
We report state-of-the-art results for text classification task on several benchmark datasets.
arXiv Detail & Related papers (2020-09-08T21:55:22Z) - Unpaired Multi-modal Segmentation via Knowledge Distillation [77.39798870702174]
We propose a novel learning scheme for unpaired cross-modality image segmentation.
In our method, we heavily reuse network parameters, by sharing all convolutional kernels across CT and MRI.
We have extensively validated our approach on two multi-class segmentation problems.
arXiv Detail & Related papers (2020-01-06T20:03:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.