Semantic Segmentation In-the-Wild Without Seeing Any Segmentation
Examples
- URL: http://arxiv.org/abs/2112.03185v1
- Date: Mon, 6 Dec 2021 17:32:38 GMT
- Title: Semantic Segmentation In-the-Wild Without Seeing Any Segmentation
Examples
- Authors: Nir Zabari, Yedid Hoshen
- Abstract summary: We propose a novel approach for creating semantic segmentation masks for every object.
Our method takes as input the image-level labels of the class categories present in the image.
The output of this stage provides pixel-level pseudo-labels, instead of the manual pixel-level labels required by supervised methods.
- Score: 34.97652735163338
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Semantic segmentation is a key computer vision task that has been actively
researched for decades. In recent years, supervised methods have reached
unprecedented accuracy, however they require many pixel-level annotations for
every new class category which is very time-consuming and expensive.
Additionally, the ability of current semantic segmentation networks to handle a
large number of categories is limited. That means that images containing rare
class categories are unlikely to be well segmented by current methods. In this
paper we propose a novel approach for creating semantic segmentation masks for
every object, without the need for training segmentation networks or seeing any
segmentation masks. Our method takes as input the image-level labels of the
class categories present in the image; they can be obtained automatically or
manually. We utilize a vision-language embedding model (specifically CLIP) to
create a rough segmentation map for each class, using model interpretability
methods. We refine the maps using a test-time augmentation technique. The
output of this stage provides pixel-level pseudo-labels, instead of the manual
pixel-level labels required by supervised methods. Given the pseudo-labels, we
utilize single-image segmentation techniques to obtain high-quality output
segmentation masks. Our method is shown quantitatively and qualitatively to
outperform methods that use a similar amount of supervision. Our results are
particularly remarkable for images containing rare categories.
Related papers
- Unsupervised Universal Image Segmentation [59.0383635597103]
We propose an Unsupervised Universal model (U2Seg) adept at performing various image segmentation tasks.
U2Seg generates pseudo semantic labels for these segmentation tasks via leveraging self-supervised models.
We then self-train the model on these pseudo semantic labels, yielding substantial performance gains.
arXiv Detail & Related papers (2023-12-28T18:59:04Z) - Learning Semantic Segmentation with Query Points Supervision on Aerial Images [57.09251327650334]
We present a weakly supervised learning algorithm to train semantic segmentation algorithms.
Our proposed approach performs accurate semantic segmentation and improves efficiency by significantly reducing the cost and time required for manual annotation.
arXiv Detail & Related papers (2023-09-11T14:32:04Z) - Synthetic Instance Segmentation from Semantic Image Segmentation Masks [15.477053085267404]
We propose a novel paradigm called Synthetic Instance (SISeg)
SISeg instance segmentation results by leveraging image masks generated by existing semantic segmentation models.
In other words, the proposed model does not need extra manpower or higher computational expenses.
arXiv Detail & Related papers (2023-08-02T05:13:02Z) - Iterative Few-shot Semantic Segmentation from Image Label Text [36.53926941601841]
Few-shot semantic segmentation aims to learn to segment unseen class objects with the guidance of only a few support images.
We propose a general framework to generate coarse masks with the help of the powerful vision-language model CLIP.
Our method owns an excellent generalization ability for the images in the wild and uncommon classes.
arXiv Detail & Related papers (2023-03-10T01:48:14Z) - Per-Pixel Classification is Not All You Need for Semantic Segmentation [184.2905747595058]
Mask classification is sufficiently general to solve both semantic- and instance-level segmentation tasks.
We propose MaskFormer, a simple mask classification model which predicts a set of binary masks.
Our method outperforms both current state-of-the-art semantic (55.6 mIoU on ADE20K) and panoptic segmentation (52.7 PQ on COCO) models.
arXiv Detail & Related papers (2021-07-13T17:59:50Z) - Segmenter: Transformer for Semantic Segmentation [79.9887988699159]
We introduce Segmenter, a transformer model for semantic segmentation.
We build on the recent Vision Transformer (ViT) and extend it to semantic segmentation.
It outperforms the state of the art on the challenging ADE20K dataset and performs on-par on Pascal Context and Cityscapes.
arXiv Detail & Related papers (2021-05-12T13:01:44Z) - A Closer Look at Self-training for Zero-Label Semantic Segmentation [53.4488444382874]
Being able to segment unseen classes not observed during training is an important technical challenge in deep learning.
Prior zero-label semantic segmentation works approach this task by learning visual-semantic embeddings or generative models.
We propose a consistency regularizer to filter out noisy pseudo-labels by taking the intersections of the pseudo-labels generated from different augmentations of the same image.
arXiv Detail & Related papers (2021-04-21T14:34:33Z) - Semantically Meaningful Class Prototype Learning for One-Shot Image
Semantic Segmentation [58.96902899546075]
One-shot semantic image segmentation aims to segment the object regions for the novel class with only one annotated image.
Recent works adopt the episodic training strategy to mimic the expected situation at testing time.
We propose to leverage the multi-class label information during the episodic training. It will encourage the network to generate more semantically meaningful features for each category.
arXiv Detail & Related papers (2021-02-22T12:07:35Z) - PCAMs: Weakly Supervised Semantic Segmentation Using Point Supervision [12.284208932393073]
This paper presents a novel procedure for producing semantic segmentation from images given some point level annotations.
We propose training a CNN that is normally fully supervised using our pseudo labels in place of ground truth labels.
Our method achieves state of the art results for point supervised semantic segmentation on the PASCAL VOC 2012 dataset citeeveringham2010pascal, even outperforming state of the art methods for stronger bounding box and squiggle supervision.
arXiv Detail & Related papers (2020-07-10T21:25:27Z) - Discovering Latent Classes for Semi-Supervised Semantic Segmentation [18.5909667833129]
This paper studies the problem of semi-supervised semantic segmentation.
We learn latent classes consistent with semantic classes on labeled images.
We show that the proposed method achieves state of the art results for semi-supervised semantic segmentation.
arXiv Detail & Related papers (2019-12-30T14:16:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.