Maximize the Exploration of Congeneric Semantics for Weakly Supervised
Semantic Segmentation
- URL: http://arxiv.org/abs/2110.03982v1
- Date: Fri, 8 Oct 2021 08:59:16 GMT
- Title: Maximize the Exploration of Congeneric Semantics for Weakly Supervised
Semantic Segmentation
- Authors: Ke Zhang, Sihong Chen, Qi Ju, Yong Jiang, Yucong Li, Xin He
- Abstract summary: We construct a graph neural network (P-GNN) based on the self-detected patches from different images that contain the same class labels.
We conduct experiments on the popular PASCAL VOC 2012 benchmarks, and our model yields state-of-the-art performance.
- Score: 27.155133686127474
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: With the increase in the number of image data and the lack of corresponding
labels, weakly supervised learning has drawn a lot of attention recently in
computer vision tasks, especially in the fine-grained semantic segmentation
problem. To alleviate human efforts from expensive pixel-by-pixel annotations,
our method focuses on weakly supervised semantic segmentation (WSSS) with
image-level tags, which are much easier to obtain. As a huge gap exists between
pixel-level segmentation and image-level labels, how to reflect the image-level
semantic information on each pixel is an important question. To explore the
congeneric semantic regions from the same class to the maximum, we construct
the patch-level graph neural network (P-GNN) based on the self-detected patches
from different images that contain the same class labels. Patches can frame the
objects as much as possible and include as little background as possible. The
graph network that is established with patches as the nodes can maximize the
mutual learning of similar objects. We regard the embedding vectors of patches
as nodes, and use transformer-based complementary learning module to construct
weighted edges according to the embedding similarity between different nodes.
Moreover, to better supplement semantic information, we propose
soft-complementary loss functions matched with the whole network structure. We
conduct experiments on the popular PASCAL VOC 2012 benchmarks, and our model
yields state-of-the-art performance.
Related papers
- Towards Open-Vocabulary Semantic Segmentation Without Semantic Labels [53.8817160001038]
We propose a novel method, PixelCLIP, to adapt the CLIP image encoder for pixel-level understanding.
To address the challenges of leveraging masks without semantic labels, we devise an online clustering algorithm.
PixelCLIP shows significant performance improvements over CLIP and competitive results compared to caption-supervised methods.
arXiv Detail & Related papers (2024-09-30T01:13:03Z) - Superpixel Semantics Representation and Pre-training for Vision-Language Task [11.029236633301222]
coarse-grained semantic interactions in image space should not be ignored.
This paper proposes superpixels as comprehensive and robust visual primitives.
It allows parsing the entire image as a fine-to-coarse visual hierarchy.
arXiv Detail & Related papers (2023-10-20T12:26:04Z) - Learning Semantic Segmentation with Query Points Supervision on Aerial Images [57.09251327650334]
We present a weakly supervised learning algorithm to train semantic segmentation algorithms.
Our proposed approach performs accurate semantic segmentation and improves efficiency by significantly reducing the cost and time required for manual annotation.
arXiv Detail & Related papers (2023-09-11T14:32:04Z) - A Pixel-Level Meta-Learner for Weakly Supervised Few-Shot Semantic
Segmentation [40.27705176115985]
Few-shot semantic segmentation addresses the learning task in which only few images with ground truth pixel-level labels are available for the novel classes of interest.
We propose a novel meta-learning framework, which predicts pseudo pixel-level segmentation masks from a limited amount of data and their semantic labels.
Our proposed learning model can be viewed as a pixel-level meta-learner.
arXiv Detail & Related papers (2021-11-02T08:28:11Z) - Semantic Segmentation with Generative Models: Semi-Supervised Learning
and Strong Out-of-Domain Generalization [112.68171734288237]
We propose a novel framework for discriminative pixel-level tasks using a generative model of both images and labels.
We learn a generative adversarial network that captures the joint image-label distribution and is trained efficiently using a large set of unlabeled images.
We demonstrate strong in-domain performance compared to several baselines, and are the first to showcase extreme out-of-domain generalization.
arXiv Detail & Related papers (2021-04-12T21:41:25Z) - Exploring Cross-Image Pixel Contrast for Semantic Segmentation [130.22216825377618]
We propose a pixel-wise contrastive framework for semantic segmentation in the fully supervised setting.
The core idea is to enforce pixel embeddings belonging to a same semantic class to be more similar than embeddings from different classes.
Our method can be effortlessly incorporated into existing segmentation frameworks without extra overhead during testing.
arXiv Detail & Related papers (2021-01-28T11:35:32Z) - Group-Wise Semantic Mining for Weakly Supervised Semantic Segmentation [49.90178055521207]
This work addresses weakly supervised semantic segmentation (WSSS), with the goal of bridging the gap between image-level annotations and pixel-level segmentation.
We formulate WSSS as a novel group-wise learning task that explicitly models semantic dependencies in a group of images to estimate more reliable pseudo ground-truths.
In particular, we devise a graph neural network (GNN) for group-wise semantic mining, wherein input images are represented as graph nodes.
arXiv Detail & Related papers (2020-12-09T12:40:13Z) - Mining Cross-Image Semantics for Weakly Supervised Semantic Segmentation [128.03739769844736]
Two neural co-attentions are incorporated into the classifier to capture cross-image semantic similarities and differences.
In addition to boosting object pattern learning, the co-attention can leverage context from other related images to improve localization map inference.
Our algorithm sets new state-of-the-arts on all these settings, demonstrating well its efficacy and generalizability.
arXiv Detail & Related papers (2020-07-03T21:53:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.