Fully Convolutional Networks for Panoptic Segmentation with Point-based
Supervision
- URL: http://arxiv.org/abs/2108.07682v2
- Date: Wed, 18 Aug 2021 02:15:32 GMT
- Title: Fully Convolutional Networks for Panoptic Segmentation with Point-based
Supervision
- Authors: Yanwei Li, Hengshuang Zhao, Xiaojuan Qi, Yukang Chen, Lu Qi, Liwei
Wang, Zeming Li, Jian Sun, Jiaya Jia
- Abstract summary: We present a conceptually simple, strong, and efficient framework for fully- and weakly-supervised panoptic segmentation, called Panoptic FCN.
Our approach aims to represent and predict foreground things and background stuff in a unified fully convolutional pipeline.
Panoptic FCN encodes each object instance or stuff category with the proposed kernel generator and produces the prediction by convolving the high-resolution feature directly.
- Score: 88.71403886207071
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we present a conceptually simple, strong, and efficient
framework for fully- and weakly-supervised panoptic segmentation, called
Panoptic FCN. Our approach aims to represent and predict foreground things and
background stuff in a unified fully convolutional pipeline, which can be
optimized with point-based fully or weak supervision. In particular, Panoptic
FCN encodes each object instance or stuff category with the proposed kernel
generator and produces the prediction by convolving the high-resolution feature
directly. With this approach, instance-aware and semantically consistent
properties for things and stuff can be respectively satisfied in a simple
generate-kernel-then-segment workflow. Without extra boxes for localization or
instance separation, the proposed approach outperforms the previous box-based
and -free models with high efficiency. Furthermore, we propose a new form of
point-based annotation for weakly-supervised panoptic segmentation. It only
needs several random points for both things and stuff, which dramatically
reduces the annotation cost of human. The proposed Panoptic FCN is also proved
to have much superior performance in this weakly-supervised setting, which
achieves 82% of the fully-supervised performance with only 20 randomly
annotated points per instance. Extensive experiments demonstrate the
effectiveness and efficiency of Panoptic FCN on COCO, VOC 2012, Cityscapes, and
Mapillary Vistas datasets. And it sets up a new leading benchmark for both
fully- and weakly-supervised panoptic segmentation. Our code and models are
made publicly available at https://github.com/dvlab-research/PanopticFCN
Related papers
- Efficient and Effective Deep Multi-view Subspace Clustering [9.6753782215283]
We propose a novel deep framework, termed Efficient and Effective deep Multi-View Subspace Clustering (E$2$MVSC)
Instead of a parameterized FC layer, we design a Relation-Metric Net that decouples network parameter scale from sample numbers for greater computational efficiency.
E$2$MVSC yields comparable results to existing methods and achieves state-of-the-art performance in various types of multi-view datasets.
arXiv Detail & Related papers (2023-10-15T03:08:25Z) - Pointly-Supervised Panoptic Segmentation [106.68888377104886]
We propose a new approach to applying point-level annotations for weakly-supervised panoptic segmentation.
Instead of the dense pixel-level labels used by fully supervised methods, point-level labels only provide a single point for each target as supervision.
We formulate the problem in an end-to-end framework by simultaneously generating panoptic pseudo-masks from point-level labels and learning from them.
arXiv Detail & Related papers (2022-10-25T12:03:51Z) - Affinity Attention Graph Neural Network for Weakly Supervised Semantic
Segmentation [86.44301443789763]
We propose Affinity Attention Graph Neural Network ($A2$GNN) for weakly supervised semantic segmentation.
Our approach achieves new state-of-the-art performances on Pascal VOC 2012 datasets.
arXiv Detail & Related papers (2021-06-08T02:19:21Z) - Exemplar-Based Open-Set Panoptic Segmentation Network [79.99748041746592]
We extend panoptic segmentation to the open-world and introduce an open-set panoptic segmentation (OPS) task.
We investigate the practical challenges of the task and construct a benchmark on top of an existing dataset, COCO.
We propose a novel exemplar-based open-set panoptic segmentation network (EOPSN) inspired by exemplar theory.
arXiv Detail & Related papers (2021-05-18T07:59:21Z) - Panoptic-PolarNet: Proposal-free LiDAR Point Cloud Panoptic Segmentation [21.296942497092402]
We present a fast and robust LiDAR point cloud panoptic segmentation framework, referred to as Panoptic-PolarNet.
We learn both semantic segmentation and class-agnostic instance clustering in a single inference network using a polar Bird's Eye View representation.
Our experiments show that Panoptic-PolarNet outperforms the baseline methods on Semantic KITTI and nuScenes datasets with an almost real-time inference speed.
arXiv Detail & Related papers (2021-03-27T18:31:40Z) - Fully Convolutional Networks for Panoptic Segmentation [91.84686839549488]
We present a conceptually simple, strong, and efficient framework for panoptic segmentation, called Panoptic FCN.
Our approach aims to represent and predict foreground things and background stuff in a unified fully convolutional pipeline.
Panoptic FCN encodes each object instance or stuff category into a specific kernel weight with the proposed kernel generator.
arXiv Detail & Related papers (2020-12-01T18:31:41Z) - Towards Efficient Scene Understanding via Squeeze Reasoning [71.1139549949694]
We propose a novel framework called Squeeze Reasoning.
Instead of propagating information on the spatial map, we first learn to squeeze the input feature into a channel-wise global vector.
We show that our approach can be modularized as an end-to-end trained block and can be easily plugged into existing networks.
arXiv Detail & Related papers (2020-11-06T12:17:01Z) - Towards Bounding-Box Free Panoptic Segmentation [16.4548904544277]
We introduce a new Bounding-Box Free Network (BBFNet) for panoptic segmentation.
BBFNet predicts coarse watershed levels and uses them to detect large instance candidates where boundaries are well defined.
For smaller instances, whose boundaries are less reliable, BBFNet also predicts instance centers by means of Hough voting followed by mean-shift to reliably detect small objects.
arXiv Detail & Related papers (2020-02-18T16:34:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.