Universal Weakly Supervised Segmentation by Pixel-to-Segment Contrastive
Learning
- URL: http://arxiv.org/abs/2105.00957v1
- Date: Mon, 3 May 2021 15:49:01 GMT
- Title: Universal Weakly Supervised Segmentation by Pixel-to-Segment Contrastive
Learning
- Authors: Tsung-Wei Ke, Jyh-Jing Hwang, Stella X. Yu
- Abstract summary: We formulate weakly supervised segmentation as a semi-supervised metric learning problem.
We propose 4 types of contrastive relationships between pixels and segments in the feature space.
We deliver a universal weakly supervised segmenter with significant gains on Pascal VOC and DensePose.
- Score: 28.498782661888775
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Weakly supervised segmentation requires assigning a label to every pixel
based on training instances with partial annotations such as image-level tags,
object bounding boxes, labeled points and scribbles. This task is challenging,
as coarse annotations (tags, boxes) lack precise pixel localization whereas
sparse annotations (points, scribbles) lack broad region coverage. Existing
methods tackle these two types of weak supervision differently: Class
activation maps are used to localize coarse labels and iteratively refine the
segmentation model, whereas conditional random fields are used to propagate
sparse labels to the entire image.
We formulate weakly supervised segmentation as a semi-supervised metric
learning problem, where pixels of the same (different) semantics need to be
mapped to the same (distinctive) features. We propose 4 types of contrastive
relationships between pixels and segments in the feature space, capturing
low-level image similarity, semantic annotation, co-occurrence, and feature
affinity They act as priors; the pixel-wise feature can be learned from
training images with any partial annotations in a data-driven fashion. In
particular, unlabeled pixels in training images participate not only in
data-driven grouping within each image, but also in discriminative feature
learning within and across images. We deliver a universal weakly supervised
segmenter with significant gains on Pascal VOC and DensePose.
Related papers
- Learning Semantic Segmentation with Query Points Supervision on Aerial Images [57.09251327650334]
We present a weakly supervised learning algorithm to train semantic segmentation algorithms.
Our proposed approach performs accurate semantic segmentation and improves efficiency by significantly reducing the cost and time required for manual annotation.
arXiv Detail & Related papers (2023-09-11T14:32:04Z) - Language-driven Semantic Segmentation [88.21498323896475]
We present LSeg, a novel model for language-driven semantic image segmentation.
We use a text encoder to compute embeddings of descriptive input labels.
The encoder is trained with a contrastive objective to align pixel embeddings to the text embedding of the corresponding semantic class.
arXiv Detail & Related papers (2022-01-10T18:59:10Z) - Maximize the Exploration of Congeneric Semantics for Weakly Supervised
Semantic Segmentation [27.155133686127474]
We construct a graph neural network (P-GNN) based on the self-detected patches from different images that contain the same class labels.
We conduct experiments on the popular PASCAL VOC 2012 benchmarks, and our model yields state-of-the-art performance.
arXiv Detail & Related papers (2021-10-08T08:59:16Z) - Mixed Supervision Learning for Whole Slide Image Classification [88.31842052998319]
We propose a mixed supervision learning framework for super high-resolution images.
During the patch training stage, this framework can make use of coarse image-level labels to refine self-supervised learning.
A comprehensive strategy is proposed to suppress pixel-level false positives and false negatives.
arXiv Detail & Related papers (2021-07-02T09:46:06Z) - KRADA: Known-region-aware Domain Alignment for Open World Semantic
Segmentation [64.03817806316903]
In semantic segmentation, we aim to train a pixel-level classifier to assign category labels to all pixels in an image.
In an open world, the unlabeled test images probably contain unknown categories and have different distributions from the labeled images.
We propose an end-to-end learning framework, known-region-aware domain alignment (KRADA), to distinguish unknown classes while aligning distributions of known classes in labeled and unlabeled open-world images.
arXiv Detail & Related papers (2021-06-11T08:43:59Z) - A Closer Look at Self-training for Zero-Label Semantic Segmentation [53.4488444382874]
Being able to segment unseen classes not observed during training is an important technical challenge in deep learning.
Prior zero-label semantic segmentation works approach this task by learning visual-semantic embeddings or generative models.
We propose a consistency regularizer to filter out noisy pseudo-labels by taking the intersections of the pseudo-labels generated from different augmentations of the same image.
arXiv Detail & Related papers (2021-04-21T14:34:33Z) - Exploring Cross-Image Pixel Contrast for Semantic Segmentation [130.22216825377618]
We propose a pixel-wise contrastive framework for semantic segmentation in the fully supervised setting.
The core idea is to enforce pixel embeddings belonging to a same semantic class to be more similar than embeddings from different classes.
Our method can be effortlessly incorporated into existing segmentation frameworks without extra overhead during testing.
arXiv Detail & Related papers (2021-01-28T11:35:32Z) - From Pixel to Patch: Synthesize Context-aware Features for Zero-shot
Semantic Segmentation [22.88452754438478]
We focus on zero-shot semantic segmentation, which aims to segment unseen objects with only category-level semantic representations.
We propose a novel Context-aware feature Generation Network (CaGNet), which can synthesize context-aware pixel-wise visual features for unseen categories.
Experimental results on Pascal-VOC, Pascal-Context, and COCO-stuff show that our method significantly outperforms the existing zero-shot semantic segmentation methods.
arXiv Detail & Related papers (2020-09-25T13:26:30Z) - PCAMs: Weakly Supervised Semantic Segmentation Using Point Supervision [12.284208932393073]
This paper presents a novel procedure for producing semantic segmentation from images given some point level annotations.
We propose training a CNN that is normally fully supervised using our pseudo labels in place of ground truth labels.
Our method achieves state of the art results for point supervised semantic segmentation on the PASCAL VOC 2012 dataset citeeveringham2010pascal, even outperforming state of the art methods for stronger bounding box and squiggle supervision.
arXiv Detail & Related papers (2020-07-10T21:25:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.