One Weird Trick to Improve Your Semi-Weakly Supervised Semantic
Segmentation Model
- URL: http://arxiv.org/abs/2205.01233v1
- Date: Mon, 2 May 2022 21:46:41 GMT
- Title: One Weird Trick to Improve Your Semi-Weakly Supervised Semantic
Segmentation Model
- Authors: Wonho Bae, Junhyug Noh, Milad Jalali Asadabadi, Danica J. Sutherland
- Abstract summary: Semi-weakly supervised semantic segmentation (SWSSS) aims to train a model to identify objects in images based on a small number of images with pixel-level labels, and many more images with only image-level labels.
Most existing SWSSS algorithms extract pixel-level pseudo-labels from an image classifier - a very difficult task to do well.
We propose a method called prediction filtering, which instead of extracting pseudo-labels, just uses the classifier as a classifier.
Adding this simple post-processing method to baselines gives results competitive with or better than prior SWSSS algorithms.
- Score: 8.388356030608886
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Semi-weakly supervised semantic segmentation (SWSSS) aims to train a model to
identify objects in images based on a small number of images with pixel-level
labels, and many more images with only image-level labels. Most existing SWSSS
algorithms extract pixel-level pseudo-labels from an image classifier - a very
difficult task to do well, hence requiring complicated architectures and
extensive hyperparameter tuning on fully-supervised validation sets. We propose
a method called prediction filtering, which instead of extracting
pseudo-labels, just uses the classifier as a classifier: it ignores any
segmentation predictions from classes which the classifier is confident are not
present. Adding this simple post-processing method to baselines gives results
competitive with or better than prior SWSSS algorithms. Moreover, it is
compatible with pseudo-label methods: adding prediction filtering to existing
SWSSS algorithms further improves segmentation performance.
Related papers
- Towards Open-Vocabulary Semantic Segmentation Without Semantic Labels [53.8817160001038]
We propose a novel method, PixelCLIP, to adapt the CLIP image encoder for pixel-level understanding.
To address the challenges of leveraging masks without semantic labels, we devise an online clustering algorithm.
PixelCLIP shows significant performance improvements over CLIP and competitive results compared to caption-supervised methods.
arXiv Detail & Related papers (2024-09-30T01:13:03Z) - Exploring Simple Open-Vocabulary Semantic Segmentation [7.245983878396646]
Open-vocabulary semantic segmentation models aim to accurately assign a semantic label to each pixel in an image from a set of arbitrary open-vocabulary texts.
In this paper, we introduce S-Seg, a novel model that can achieve surprisingly strong performance without depending on any of the above elements.
arXiv Detail & Related papers (2024-01-22T18:59:29Z) - Learning Semantic Segmentation with Query Points Supervision on Aerial Images [57.09251327650334]
We present a weakly supervised learning algorithm to train semantic segmentation algorithms.
Our proposed approach performs accurate semantic segmentation and improves efficiency by significantly reducing the cost and time required for manual annotation.
arXiv Detail & Related papers (2023-09-11T14:32:04Z) - Multi-Granularity Denoising and Bidirectional Alignment for Weakly
Supervised Semantic Segmentation [75.32213865436442]
We propose an end-to-end multi-granularity denoising and bidirectional alignment (MDBA) model to alleviate the noisy label and multi-class generalization issues.
The MDBA model can reach the mIoU of 69.5% and 70.2% on validation and test sets for the PASCAL VOC 2012 dataset.
arXiv Detail & Related papers (2023-05-09T03:33:43Z) - High-fidelity Pseudo-labels for Boosting Weakly-Supervised Segmentation [17.804090651425955]
Image-level weakly-supervised segmentation (WSSS) reduces the usually vast data annotation cost by surrogate segmentation masks during training.
Our work is based on two techniques for improving CAMs; importance sampling, which is a substitute for GAP, and the feature similarity loss.
We reformulate both techniques based on binomial posteriors of multiple independent binary problems.
This has two benefits; their performance is improved and they become more general, resulting in an add-on method that can boost virtually any WSSS method.
arXiv Detail & Related papers (2023-04-05T17:43:57Z) - ISLE: A Framework for Image Level Semantic Segmentation Ensemble [5.137284292672375]
Conventional semantic segmentation networks require massive pixel-wise annotated labels to reach state-of-the-art prediction quality.
We propose ISLE, which employs an ensemble of the "pseudo-labels" for a given set of different semantic segmentation techniques on a class-wise level.
We reach up to 2.4% improvement over ISLE's individual components.
arXiv Detail & Related papers (2023-03-14T13:36:36Z) - Location-Aware Self-Supervised Transformers [74.76585889813207]
We propose to pretrain networks for semantic segmentation by predicting the relative location of image parts.
We control the difficulty of the task by masking a subset of the reference patch features visible to those of the query.
Our experiments show that this location-aware pretraining leads to representations that transfer competitively to several challenging semantic segmentation benchmarks.
arXiv Detail & Related papers (2022-12-05T16:24:29Z) - Learning to Annotate Part Segmentation with Gradient Matching [58.100715754135685]
This paper focuses on tackling semi-supervised part segmentation tasks by generating high-quality images with a pre-trained GAN.
In particular, we formulate the annotator learning as a learning-to-learn problem.
We show that our method can learn annotators from a broad range of labelled images including real images, generated images, and even analytically rendered images.
arXiv Detail & Related papers (2022-11-06T01:29:22Z) - Semantic Distribution-aware Contrastive Adaptation for Semantic
Segmentation [50.621269117524925]
Domain adaptive semantic segmentation refers to making predictions on a certain target domain with only annotations of a specific source domain.
We present a semantic distribution-aware contrastive adaptation algorithm that enables pixel-wise representation alignment.
We evaluate SDCA on multiple benchmarks, achieving considerable improvements over existing algorithms.
arXiv Detail & Related papers (2021-05-11T13:21:25Z) - Unsupervised Image Segmentation using Mutual Mean-Teaching [12.784209596867495]
We propose an unsupervised image segmentation model based on the Mutual Mean-Teaching (MMT) framework to produce more stable results.
Experimental results demonstrate that the proposed model is able to segment various types of images and achieves better performance than the existing methods.
arXiv Detail & Related papers (2020-12-16T13:13:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.