Object Segmentation Without Labels with Large-Scale Generative Models
- URL: http://arxiv.org/abs/2006.04988v2
- Date: Fri, 11 Jun 2021 09:49:40 GMT
- Title: Object Segmentation Without Labels with Large-Scale Generative Models
- Authors: Andrey Voynov, Stanislav Morozov, Artem Babenko
- Abstract summary: Recent rise of unsupervised and self-supervised learning has dramatically reduced the dependency on labeled data.
Large-scale unsupervised models can also perform a more challenging object segmentation task, requiring neither pixel-level nor image-level labeling.
We show that recent unsupervised GANs allow to differentiate between foreground/background pixels, providing high-quality saliency masks.
- Score: 43.679717400251924
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The recent rise of unsupervised and self-supervised learning has dramatically
reduced the dependency on labeled data, providing effective image
representations for transfer to downstream vision tasks. Furthermore, recent
works employed these representations in a fully unsupervised setup for image
classification, reducing the need for human labels on the fine-tuning stage as
well. This work demonstrates that large-scale unsupervised models can also
perform a more challenging object segmentation task, requiring neither
pixel-level nor image-level labeling. Namely, we show that recent unsupervised
GANs allow to differentiate between foreground/background pixels, providing
high-quality saliency masks. By extensive comparison on standard benchmarks, we
outperform existing unsupervised alternatives for object segmentation,
achieving new state-of-the-art.
Related papers
- Unsupervised Camouflaged Object Segmentation as Domain Adaptation [5.304455190825436]
In this work, we investigate a new task, namely unsupervised camouflaged object segmentation (UCOS), where the target objects own a common attribute, i.e., camouflage.
We find that the state-of-the-art unsupervised models struggle in adapting UCOS, due to the domain gap between the properties of generic and camouflaged objects.
We formulate the UCOS as a source-free unsupervised domain adaptation task (UCOS-DA), where both source labels and target labels are absent during the whole model training process.
arXiv Detail & Related papers (2023-08-08T18:46:16Z) - Distilling Self-Supervised Vision Transformers for Weakly-Supervised
Few-Shot Classification & Segmentation [58.03255076119459]
We address the task of weakly-supervised few-shot image classification and segmentation, by leveraging a Vision Transformer (ViT)
Our proposed method takes token representations from the self-supervised ViT and leverages their correlations, via self-attention, to produce classification and segmentation predictions.
Experiments on Pascal-5i and COCO-20i demonstrate significant performance gains in a variety of supervision settings.
arXiv Detail & Related papers (2023-07-07T06:16:43Z) - RaSP: Relation-aware Semantic Prior for Weakly Supervised Incremental
Segmentation [28.02204928717511]
We propose a weakly supervised approach to transfer objectness prior from the previously learned classes into the new ones.
We show how even a simple pairwise interaction between classes can significantly improve the segmentation mask quality of both old and new classes.
arXiv Detail & Related papers (2023-05-31T14:14:21Z) - Pointly-Supervised Panoptic Segmentation [106.68888377104886]
We propose a new approach to applying point-level annotations for weakly-supervised panoptic segmentation.
Instead of the dense pixel-level labels used by fully supervised methods, point-level labels only provide a single point for each target as supervision.
We formulate the problem in an end-to-end framework by simultaneously generating panoptic pseudo-masks from point-level labels and learning from them.
arXiv Detail & Related papers (2022-10-25T12:03:51Z) - Masked Unsupervised Self-training for Zero-shot Image Classification [98.23094305347709]
Masked Unsupervised Self-Training (MUST) is a new approach which leverages two different and complimentary sources of supervision: pseudo-labels and raw images.
MUST improves upon CLIP by a large margin and narrows the performance gap between unsupervised and supervised classification.
arXiv Detail & Related papers (2022-06-07T02:03:06Z) - A Pixel-Level Meta-Learner for Weakly Supervised Few-Shot Semantic
Segmentation [40.27705176115985]
Few-shot semantic segmentation addresses the learning task in which only few images with ground truth pixel-level labels are available for the novel classes of interest.
We propose a novel meta-learning framework, which predicts pseudo pixel-level segmentation masks from a limited amount of data and their semantic labels.
Our proposed learning model can be viewed as a pixel-level meta-learner.
arXiv Detail & Related papers (2021-11-02T08:28:11Z) - Semi-supervised Semantic Segmentation with Directional Context-aware
Consistency [66.49995436833667]
We focus on the semi-supervised segmentation problem where only a small set of labeled data is provided with a much larger collection of totally unlabeled images.
A preferred high-level representation should capture the contextual information while not losing self-awareness.
We present the Directional Contrastive Loss (DC Loss) to accomplish the consistency in a pixel-to-pixel manner.
arXiv Detail & Related papers (2021-06-27T03:42:40Z) - Unsupervised data augmentation for object detection [13.465808931940595]
We propose a framework making use of Generative Adversarial Networks(GAN) to perform unsupervised data augmentation.
Based on the recently supreme performance of YOLOv4, we propose a two-step pipeline that enables us to generate an image where the object lies in a certain position.
arXiv Detail & Related papers (2021-04-30T13:02:42Z) - A Closer Look at Self-training for Zero-Label Semantic Segmentation [53.4488444382874]
Being able to segment unseen classes not observed during training is an important technical challenge in deep learning.
Prior zero-label semantic segmentation works approach this task by learning visual-semantic embeddings or generative models.
We propose a consistency regularizer to filter out noisy pseudo-labels by taking the intersections of the pseudo-labels generated from different augmentations of the same image.
arXiv Detail & Related papers (2021-04-21T14:34:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.