Incremental Learning Meets Transfer Learning: Application to Multi-site
Prostate MRI Segmentation
- URL: http://arxiv.org/abs/2206.01369v1
- Date: Fri, 3 Jun 2022 02:32:01 GMT
- Title: Incremental Learning Meets Transfer Learning: Application to Multi-site
Prostate MRI Segmentation
- Authors: Chenyu You, Jinlin Xiang, Kun Su, Xiaoran Zhang, Siyuan Dong, John
Onofrey, Lawrence Staib, James S. Duncan
- Abstract summary: We propose a novel multi-site segmentation framework called incremental-transfer learning (ITL)
ITL learns a model from multi-site datasets in an end-to-end sequential fashion.
We show for the first time that leveraging our ITL training scheme is able to alleviate challenging catastrophic problems in incremental learning.
- Score: 16.50535949349874
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Many medical datasets have recently been created for medical image
segmentation tasks, and it is natural to question whether we can use them to
sequentially train a single model that (1) performs better on all these
datasets, and (2) generalizes well and transfers better to the unknown target
site domain. Prior works have achieved this goal by jointly training one model
on multi-site datasets, which achieve competitive performance on average but
such methods rely on the assumption about the availability of all training
data, thus limiting its effectiveness in practical deployment. In this paper,
we propose a novel multi-site segmentation framework called
incremental-transfer learning (ITL), which learns a model from multi-site
datasets in an end-to-end sequential fashion. Specifically, "incremental"
refers to training sequentially constructed datasets, and "transfer" is
achieved by leveraging useful information from the linear combination of
embedding features on each dataset. In addition, we introduce our ITL
framework, where we train the network including a site-agnostic encoder with
pre-trained weights and at most two segmentation decoder heads. We also design
a novel site-level incremental loss in order to generalize well on the target
domain. Second, we show for the first time that leveraging our ITL training
scheme is able to alleviate challenging catastrophic forgetting problems in
incremental learning. We conduct experiments using five challenging benchmark
datasets to validate the effectiveness of our incremental-transfer learning
approach. Our approach makes minimal assumptions on computation resources and
domain-specific expertise, and hence constitutes a strong starting point in
multi-site medical image segmentation.
Related papers
- Motor Imagery Decoding Using Ensemble Curriculum Learning and
Collaborative Training [11.157243900163376]
Multi-subject EEG datasets present several kinds of domain shifts.
These domain shifts impede robust cross-subject generalization.
We propose a two-stage model ensemble architecture built with multiple feature extractors.
We demonstrate that our model ensembling approach combines the powers of curriculum learning and collaborative training.
arXiv Detail & Related papers (2022-11-21T13:45:44Z) - Training a universal instance segmentation network for live cell images
of various cell types and imaging modalities [10.644558286623813]
We present an attempt to train a universal segmentation network for various cell types and imaging modalities.
We modified the traditional binary training targets to include three classes for direct instance segmentation.
Our method was evaluated as the best runner up during the initial submission for the primary track, and also secured the 3rd place in an additional round of competition.
arXiv Detail & Related papers (2022-07-28T18:57:30Z) - Learning from Temporal Spatial Cubism for Cross-Dataset Skeleton-based
Action Recognition [88.34182299496074]
Action labels are only available on a source dataset, but unavailable on a target dataset in the training stage.
We utilize a self-supervision scheme to reduce the domain shift between two skeleton-based action datasets.
By segmenting and permuting temporal segments or human body parts, we design two self-supervised learning classification tasks.
arXiv Detail & Related papers (2022-07-17T07:05:39Z) - Continual Coarse-to-Fine Domain Adaptation in Semantic Segmentation [22.366638308792734]
Deep neural networks are typically trained in a single shot for a specific task and data distribution.
In real world settings both the task and the domain of application can change.
We introduce the novel task of coarse-to-fine learning of semantic segmentation architectures in presence of domain shift.
arXiv Detail & Related papers (2022-01-18T13:31:19Z) - Omni-Training for Data-Efficient Deep Learning [80.28715182095975]
Recent advances reveal that a properly pre-trained model endows an important property: transferability.
A tight combination of pre-training and meta-training cannot achieve both kinds of transferability.
This motivates the proposed Omni-Training framework towards data-efficient deep learning.
arXiv Detail & Related papers (2021-10-14T16:30:36Z) - Unsupervised Domain Adaptive Learning via Synthetic Data for Person
Re-identification [101.1886788396803]
Person re-identification (re-ID) has gained more and more attention due to its widespread applications in video surveillance.
Unfortunately, the mainstream deep learning methods still need a large quantity of labeled data to train models.
In this paper, we develop a data collector to automatically generate synthetic re-ID samples in a computer game, and construct a data labeler to simultaneously annotate them.
arXiv Detail & Related papers (2021-09-12T15:51:41Z) - Curriculum Graph Co-Teaching for Multi-Target Domain Adaptation [78.28390172958643]
We identify two key aspects that can help to alleviate multiple domain-shifts in the multi-target domain adaptation (MTDA)
We propose Curriculum Graph Co-Teaching (CGCT) that uses a dual classifier head, with one of them being a graph convolutional network (GCN) which aggregates features from similar samples across the domains.
When the domain labels are available, we propose Domain-aware Curriculum Learning (DCL), a sequential adaptation strategy that first adapts on the easier target domains, followed by the harder ones.
arXiv Detail & Related papers (2021-04-01T23:41:41Z) - Towards Accurate Knowledge Transfer via Target-awareness Representation
Disentanglement [56.40587594647692]
We propose a novel transfer learning algorithm, introducing the idea of Target-awareness REpresentation Disentanglement (TRED)
TRED disentangles the relevant knowledge with respect to the target task from the original source model and used as a regularizer during fine-tuning the target model.
Experiments on various real world datasets show that our method stably improves the standard fine-tuning by more than 2% in average.
arXiv Detail & Related papers (2020-10-16T17:45:08Z) - Dual-Teacher: Integrating Intra-domain and Inter-domain Teachers for
Annotation-efficient Cardiac Segmentation [65.81546955181781]
We propose a novel semi-supervised domain adaptation approach, namely Dual-Teacher.
The student model learns the knowledge of unlabeled target data and labeled source data by two teacher models.
We demonstrate that our approach is able to concurrently utilize unlabeled data and cross-modality data with superior performance.
arXiv Detail & Related papers (2020-07-13T10:00:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.