Urban Scene Semantic Segmentation with Low-Cost Coarse Annotation
- URL: http://arxiv.org/abs/2212.07911v1
- Date: Thu, 15 Dec 2022 15:43:42 GMT
- Title: Urban Scene Semantic Segmentation with Low-Cost Coarse Annotation
- Authors: Anurag Das, Yongqin Xian, Yang He, Zeynep Akata and Bernt Schiele
- Abstract summary: coarse annotation is a low-cost but highly effective alternative for training semantic segmentation models.
We propose a coarse-to-fine self-training framework that generates pseudo labels for unlabeled regions of coarsely annotated data.
Our method achieves a significantly better performance vs annotation cost tradeoff, yielding a comparable performance to fully annotated data with only a small fraction of the annotation budget.
- Score: 107.72926721837726
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: For best performance, today's semantic segmentation methods use large and
carefully labeled datasets, requiring expensive annotation budgets. In this
work, we show that coarse annotation is a low-cost but highly effective
alternative for training semantic segmentation models. Considering the urban
scene segmentation scenario, we leverage cheap coarse annotations for
real-world captured data, as well as synthetic data to train our model and show
competitive performance compared with finely annotated real-world data.
Specifically, we propose a coarse-to-fine self-training framework that
generates pseudo labels for unlabeled regions of the coarsely annotated data,
using synthetic data to improve predictions around the boundaries between
semantic classes, and using cross-domain data augmentation to increase
diversity. Our extensive experimental results on Cityscapes and BDD100k
datasets demonstrate that our method achieves a significantly better
performance vs annotation cost tradeoff, yielding a comparable performance to
fully annotated data with only a small fraction of the annotation budget. Also,
when used as pretraining, our framework performs better compared to the
standard fully supervised setting.
Related papers
- Scribbles for All: Benchmarking Scribble Supervised Segmentation Across Datasets [51.74296438621836]
We introduce Scribbles for All, a label and training data generation algorithm for semantic segmentation trained on scribble labels.
The main limitation of scribbles as source for weak supervision is the lack of challenging datasets for scribble segmentation.
Scribbles for All provides scribble labels for several popular segmentation datasets and provides an algorithm to automatically generate scribble labels for any dataset with dense annotations.
arXiv Detail & Related papers (2024-08-22T15:29:08Z) - A Self Supervised StyleGAN for Image Annotation and Classification with
Extremely Limited Labels [35.43549147657739]
We propose SS-StyleGAN, a self-supervised approach for image annotation and classification suitable for extremely small annotated datasets.
We show that the proposed method attains strong classification results using small labeled datasets of sizes 50 and even 10.
arXiv Detail & Related papers (2023-12-26T09:46:50Z) - ScribbleGen: Generative Data Augmentation Improves Scribble-supervised Semantic Segmentation [10.225021032417589]
We propose ScribbleGen, a generative data augmentation method for scribble-supervised semantic segmentation.
We leverage a ControlNet diffusion model conditioned on semantic scribbles to produce high-quality training data.
We show that our framework significantly improves segmentation performance on small datasets, even surpassing fully-supervised segmentation.
arXiv Detail & Related papers (2023-11-28T13:44:33Z) - Pseudo Label-Guided Data Fusion and Output Consistency for
Semi-Supervised Medical Image Segmentation [9.93871075239635]
We propose the PLGDF framework, which builds upon the mean teacher network for segmenting medical images with less annotation.
We propose a novel pseudo-label utilization scheme, which combines labeled and unlabeled data to augment the dataset effectively.
Our framework yields superior performance compared to six state-of-the-art semi-supervised learning methods.
arXiv Detail & Related papers (2023-11-17T06:36:43Z) - Full or Weak annotations? An adaptive strategy for budget-constrained
annotation campaigns [3.1318537187387787]
We propose a novel approach to determine annotation strategies for segmentation datasets.
Our method sequentially determines proportions of segmentation and classification annotations to collect for budget-fractions.
We show in our experiments that our approach yields annotations that perform very close to the optimal for a number of different annotation budgets and datasets.
arXiv Detail & Related papers (2023-03-21T08:41:54Z) - LESS: Label-Efficient Semantic Segmentation for LiDAR Point Clouds [62.49198183539889]
We propose a label-efficient semantic segmentation pipeline for outdoor scenes with LiDAR point clouds.
Our method co-designs an efficient labeling process with semi/weakly supervised learning.
Our proposed method is even highly competitive compared to the fully supervised counterpart with 100% labels.
arXiv Detail & Related papers (2022-10-14T19:13:36Z) - MSeg: A Composite Dataset for Multi-domain Semantic Segmentation [100.17755160696939]
We present MSeg, a composite dataset that unifies semantic segmentation datasets from different domains.
We reconcile the generalization and bring the pixel-level annotations into alignment by relabeling more than 220,000 object masks in more than 80,000 images.
A model trained on MSeg ranks first on the WildDash-v1 leaderboard for robust semantic segmentation, with no exposure to WildDash data during training.
arXiv Detail & Related papers (2021-12-27T16:16:35Z) - Improving Zero and Few-Shot Abstractive Summarization with Intermediate
Fine-tuning and Data Augmentation [101.26235068460551]
Models pretrained with self-supervised objectives on large text corpora achieve state-of-the-art performance on English text summarization tasks.
Models are typically fine-tuned on hundreds of thousands of data points, an infeasible requirement when applying summarization to new, niche domains.
We introduce a novel and generalizable method, called WikiTransfer, for fine-tuning pretrained models for summarization in an unsupervised, dataset-specific manner.
arXiv Detail & Related papers (2020-10-24T08:36:49Z) - Improving Semantic Segmentation via Self-Training [75.07114899941095]
We show that we can obtain state-of-the-art results using a semi-supervised approach, specifically a self-training paradigm.
We first train a teacher model on labeled data, and then generate pseudo labels on a large set of unlabeled data.
Our robust training framework can digest human-annotated and pseudo labels jointly and achieve top performances on Cityscapes, CamVid and KITTI datasets.
arXiv Detail & Related papers (2020-04-30T17:09:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.