Learning Category- and Instance-Aware Pixel Embedding for Fast Panoptic
Segmentation
- URL: http://arxiv.org/abs/2009.13342v2
- Date: Wed, 16 Jun 2021 01:13:09 GMT
- Title: Learning Category- and Instance-Aware Pixel Embedding for Fast Panoptic
Segmentation
- Authors: Naiyu Gao, Yanhu Shan, Xin Zhao, Kaiqi Huang
- Abstract summary: Panoptic segmentation (PS) is a complex scene understanding task.
PS results are simply derived by assigning each pixel to a detected instance or a stuff class.
Our method not only demonstrates fast inference speed but also the first one-stage method to achieve comparable performance to two-stage methods.
- Score: 47.26296379603003
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Panoptic segmentation (PS) is a complex scene understanding task that
requires providing high-quality segmentation for both thing objects and stuff
regions. Previous methods handle these two classes with semantic and instance
segmentation modules separately, following with heuristic fusion or additional
modules to resolve the conflicts between the two outputs. This work simplifies
this pipeline of PS by consistently modeling the two classes with a novel PS
framework, which extends a detection model with an extra module to predict
category- and instance-aware pixel embedding (CIAE). CIAE is a novel pixel-wise
embedding feature that encodes both semantic-classification and
instance-distinction information. At the inference process, PS results are
simply derived by assigning each pixel to a detected instance or a stuff class
according to the learned embedding. Our method not only demonstrates fast
inference speed but also the first one-stage method to achieve comparable
performance to two-stage methods on the challenging COCO benchmark.
Related papers
- UniParser: Multi-Human Parsing with Unified Correlation Representation Learning [18.372259148872228]
This paper introduces Uni, which integrates instance-level and category-level representations in three key aspects.
Uni achieves 49.3% AP on MHPv2.0 and 60.4% AP on CIHP.
arXiv Detail & Related papers (2023-10-13T10:03:01Z) - Masked Cross-image Encoding for Few-shot Segmentation [16.445813548503708]
Few-shot segmentation (FSS) is a dense prediction task that aims to infer the pixel-wise labels of unseen classes using only a limited number of annotated images.
We propose a joint learning method termed Masked Cross-Image MCE, which is designed to capture common visual properties that describe object details and to learn bidirectional inter-image dependencies that enhance feature interaction.
arXiv Detail & Related papers (2023-08-22T05:36:39Z) - Semantic-Aware Dual Contrastive Learning for Multi-label Image
Classification [8.387933969327852]
We propose a novel semantic-aware dual contrastive learning framework that incorporates sample-to-sample contrastive learning.
Specifically, we leverage semantic-aware representation learning to extract category-related local discriminative features.
Our proposed method is effective and outperforms the state-of-the-art methods.
arXiv Detail & Related papers (2023-07-19T01:57:31Z) - APANet: Adaptive Prototypes Alignment Network for Few-Shot Semantic
Segmentation [56.387647750094466]
Few-shot semantic segmentation aims to segment novel-class objects in a given query image with only a few labeled support images.
Most advanced solutions exploit a metric learning framework that performs segmentation through matching each query feature to a learned class-specific prototype.
We present an adaptive prototype representation by introducing class-specific and class-agnostic prototypes.
arXiv Detail & Related papers (2021-11-24T04:38:37Z) - Dual Prototypical Contrastive Learning for Few-shot Semantic
Segmentation [55.339405417090084]
We propose a dual prototypical contrastive learning approach tailored to the few-shot semantic segmentation (FSS) task.
The main idea is to encourage the prototypes more discriminative by increasing inter-class distance while reducing intra-class distance in prototype feature space.
We demonstrate that the proposed dual contrastive learning approach outperforms state-of-the-art FSS methods on PASCAL-5i and COCO-20i datasets.
arXiv Detail & Related papers (2021-11-09T08:14:50Z) - SOLO: A Simple Framework for Instance Segmentation [84.00519148562606]
"instance categories" assigns categories to each pixel within an instance according to the instance's location.
"SOLO" is a simple, direct, and fast framework for instance segmentation with strong performance.
Our approach achieves state-of-the-art results for instance segmentation in terms of both speed and accuracy.
arXiv Detail & Related papers (2021-06-30T09:56:54Z) - BoundarySqueeze: Image Segmentation as Boundary Squeezing [104.43159799559464]
We propose a novel method for fine-grained high-quality image segmentation of both objects and scenes.
Inspired by dilation and erosion from morphological image processing techniques, we treat the pixel level segmentation problems as squeezing object boundary.
Our method yields large gains on COCO, Cityscapes, for both instance and semantic segmentation and outperforms previous state-of-the-art PointRend in both accuracy and speed under the same setting.
arXiv Detail & Related papers (2021-05-25T04:58:51Z) - SCNet: Enhancing Few-Shot Semantic Segmentation by Self-Contrastive
Background Prototypes [56.387647750094466]
Few-shot semantic segmentation aims to segment novel-class objects in a query image with only a few annotated examples.
Most of advanced solutions exploit a metric learning framework that performs segmentation through matching each pixel to a learned foreground prototype.
This framework suffers from biased classification due to incomplete construction of sample pairs with the foreground prototype only.
arXiv Detail & Related papers (2021-04-19T11:21:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.