Cross-Dataset Collaborative Learning for Semantic Segmentation
- URL: http://arxiv.org/abs/2103.11351v1
- Date: Sun, 21 Mar 2021 09:59:47 GMT
- Title: Cross-Dataset Collaborative Learning for Semantic Segmentation
- Authors: Li Wang, Dong Li, Yousong Zhu, Lu Tian, Yi Shan
- Abstract summary: We present a simple, flexible, and general method for semantic segmentation, termed Cross-Dataset Collaborative Learning (CDCL)
Given multiple labeled datasets, we aim to improve the generalization and discrimination of feature representations on each dataset.
We conduct extensive evaluations on four diverse datasets, i.e., Cityscapes, BDD100K, CamVid, and COCO Stuff, with single-dataset and cross-dataset settings.
- Score: 17.55660581677053
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recent work attempts to improve semantic segmentation performance by
exploring well-designed architectures on a target dataset. However, it remains
challenging to build a unified system that simultaneously learns from various
datasets due to the inherent distribution shift across different datasets. In
this paper, we present a simple, flexible, and general method for semantic
segmentation, termed Cross-Dataset Collaborative Learning (CDCL). Given
multiple labeled datasets, we aim to improve the generalization and
discrimination of feature representations on each dataset. Specifically, we
first introduce a family of Dataset-Aware Blocks (DAB) as the fundamental
computing units of the network, which help capture homogeneous representations
and heterogeneous statistics across different datasets. Second, we propose a
Dataset Alternation Training (DAT) mechanism to efficiently facilitate the
optimization procedure. We conduct extensive evaluations on four diverse
datasets, i.e., Cityscapes, BDD100K, CamVid, and COCO Stuff, with
single-dataset and cross-dataset settings. Experimental results demonstrate our
method consistently achieves notable improvements over prior single-dataset and
cross-dataset training methods without introducing extra FLOPs. Particularly,
with the same architecture of PSPNet (ResNet-18), our method outperforms the
single-dataset baseline by 5.65\%, 6.57\%, and 5.79\% of mIoU on the validation
sets of Cityscapes, BDD100K, CamVid, respectively. Code and models will be
released.
Related papers
- A-Eval: A Benchmark for Cross-Dataset Evaluation of Abdominal
Multi-Organ Segmentation [38.644744669074775]
We introduce A-Eval, a benchmark for the cross-dataset Evaluation ('Eval') of Abdominal ('A') multi-organ segmentation.
We employ training sets from four large-scale public datasets: FLARE22, AMOS, WORD, and TotalSegmentator.
We evaluate the generalizability of various models using the A-Eval benchmark, with a focus on diverse data usage scenarios.
arXiv Detail & Related papers (2023-09-07T17:59:50Z) - infoVerse: A Universal Framework for Dataset Characterization with
Multidimensional Meta-information [68.76707843019886]
infoVerse is a universal framework for dataset characterization.
infoVerse captures multidimensional characteristics of datasets by incorporating various model-driven meta-information.
In three real-world applications (data pruning, active learning, and data annotation), the samples chosen on infoVerse space consistently outperform strong baselines.
arXiv Detail & Related papers (2023-05-30T18:12:48Z) - Quality Not Quantity: On the Interaction between Dataset Design and
Robustness of CLIP [43.7219097444333]
We introduce a testbed of six publicly available data sources to investigate how pre-training distributions induce robustness in CLIP.
We find that the performance of the pre-training data varies substantially across distribution shifts.
We find that combining multiple sources does not necessarily yield better models, but rather dilutes the robustness of the best individual data source.
arXiv Detail & Related papers (2022-08-10T18:24:23Z) - Learning from Temporal Spatial Cubism for Cross-Dataset Skeleton-based
Action Recognition [88.34182299496074]
Action labels are only available on a source dataset, but unavailable on a target dataset in the training stage.
We utilize a self-supervision scheme to reduce the domain shift between two skeleton-based action datasets.
By segmenting and permuting temporal segments or human body parts, we design two self-supervised learning classification tasks.
arXiv Detail & Related papers (2022-07-17T07:05:39Z) - Detection Hub: Unifying Object Detection Datasets via Query Adaptation
on Language Embedding [137.3719377780593]
A new design (named Detection Hub) is dataset-aware and category-aligned.
It mitigates the dataset inconsistency and provides coherent guidance for the detector to learn across multiple datasets.
The categories across datasets are semantically aligned into a unified space by replacing one-hot category representations with word embedding.
arXiv Detail & Related papers (2022-06-07T17:59:44Z) - Structurally Diverse Sampling Reduces Spurious Correlations in Semantic
Parsing Datasets [51.095144091781734]
We propose a novel algorithm for sampling a structurally diverse set of instances from a labeled instance pool with structured outputs.
We show that our algorithm performs competitively with or better than prior algorithms in not only compositional template splits but also traditional IID splits.
In general, we find that diverse train sets lead to better generalization than random training sets of the same size in 9 out of 10 dataset-split pairs.
arXiv Detail & Related papers (2022-03-16T07:41:27Z) - MSeg: A Composite Dataset for Multi-domain Semantic Segmentation [100.17755160696939]
We present MSeg, a composite dataset that unifies semantic segmentation datasets from different domains.
We reconcile the generalization and bring the pixel-level annotations into alignment by relabeling more than 220,000 object masks in more than 80,000 images.
A model trained on MSeg ranks first on the WildDash-v1 leaderboard for robust semantic segmentation, with no exposure to WildDash data during training.
arXiv Detail & Related papers (2021-12-27T16:16:35Z) - Towards Federated Bayesian Network Structure Learning with Continuous
Optimization [14.779035801521717]
We present a cross-silo federated learning approach to estimate the structure of Bayesian network.
We develop a distributed structure learning method based on continuous optimization.
arXiv Detail & Related papers (2021-10-18T14:36:05Z) - Multi-dataset Pretraining: A Unified Model for Semantic Segmentation [97.61605021985062]
We propose a unified framework, termed as Multi-Dataset Pretraining, to take full advantage of the fragmented annotations of different datasets.
This is achieved by first pretraining the network via the proposed pixel-to-prototype contrastive loss over multiple datasets.
In order to better model the relationship among images and classes from different datasets, we extend the pixel level embeddings via cross dataset mixing.
arXiv Detail & Related papers (2021-06-08T06:13:11Z) - Elastic Coupled Co-clustering for Single-Cell Genomic Data [0.0]
Single-cell technologies have enabled us to profile genomic features at unprecedented resolution.
Data integration can potentially lead to a better performance of clustering algorithms.
In this work, we formulate the problem in an unsupervised transfer learning framework.
arXiv Detail & Related papers (2020-03-29T08:21:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.