Pixel Consensus Voting for Panoptic Segmentation
- URL: http://arxiv.org/abs/2004.01849v1
- Date: Sat, 4 Apr 2020 04:33:45 GMT
- Title: Pixel Consensus Voting for Panoptic Segmentation
- Authors: Haochen Wang, Ruotian Luo, Michael Maire, Greg Shakhnarovich
- Abstract summary: Pixel Consensus Voting is a framework for instance segmentation based on the Generalized Hough transform.
We implement vote aggregation and backprojection using native operators of a convolutional neural network.
- Score: 34.76582105480241
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The core of our approach, Pixel Consensus Voting, is a framework for instance
segmentation based on the Generalized Hough transform. Pixels cast discretized,
probabilistic votes for the likely regions that contain instance centroids. At
the detected peaks that emerge in the voting heatmap, backprojection is applied
to collect pixels and produce instance masks. Unlike a sliding window detector
that densely enumerates object proposals, our method detects instances as a
result of the consensus among pixel-wise votes. We implement vote aggregation
and backprojection using native operators of a convolutional neural network.
The discretization of centroid voting reduces the training of instance
segmentation to pixel labeling, analogous and complementary to FCN-style
semantic segmentation, leading to an efficient and unified architecture that
jointly models things and stuff. We demonstrate the effectiveness of our
pipeline on COCO and Cityscapes Panoptic Segmentation and obtain competitive
results. Code will be open-sourced.
Related papers
- Pixel-Level Clustering Network for Unsupervised Image Segmentation [3.69853388955692]
We present a pixel-level clustering framework for segmenting images into regions without using ground truth annotations.
We also propose a training strategy that utilizes intra-consistency within each superpixel, inter-similarity/dissimilarity between neighboring superpixels, and structural similarity between images.
arXiv Detail & Related papers (2023-10-24T23:06:29Z) - Voting Network for Contour Levee Farmland Segmentation and
Classification [3.3675306464999357]
High-resolution aerial imagery allows fine details in the segmentation of farmlands.
We present an end-to-end trainable network for segmenting farmlands with contour levees from high-resolution aerial imagery.
arXiv Detail & Related papers (2023-09-28T16:16:08Z) - Unified Mask Embedding and Correspondence Learning for Self-Supervised
Video Segmentation [76.40565872257709]
We develop a unified framework which simultaneously models cross-frame dense correspondence for locally discriminative feature learning.
It is able to directly learn to perform mask-guided sequential segmentation from unlabeled videos.
Our algorithm sets state-of-the-arts on two standard benchmarks (i.e., DAVIS17 and YouTube-VOS)
arXiv Detail & Related papers (2023-03-17T16:23:36Z) - Which Pixel to Annotate: a Label-Efficient Nuclei Segmentation Framework [70.18084425770091]
Deep neural networks have been widely applied in nuclei instance segmentation of H&E stained pathology images.
It is inefficient and unnecessary to label all pixels for a dataset of nuclei images which usually contain similar and redundant patterns.
We propose a novel full nuclei segmentation framework that chooses only a few image patches to be annotated, augments the training set from the selected samples, and achieves nuclei segmentation in a semi-supervised manner.
arXiv Detail & Related papers (2022-12-20T14:53:26Z) - Pointly-Supervised Panoptic Segmentation [106.68888377104886]
We propose a new approach to applying point-level annotations for weakly-supervised panoptic segmentation.
Instead of the dense pixel-level labels used by fully supervised methods, point-level labels only provide a single point for each target as supervision.
We formulate the problem in an end-to-end framework by simultaneously generating panoptic pseudo-masks from point-level labels and learning from them.
arXiv Detail & Related papers (2022-10-25T12:03:51Z) - From Explanations to Segmentation: Using Explainable AI for Image
Segmentation [1.8581514902689347]
We build upon the advances of the Explainable AI (XAI) community and extract a pixel-wise binary segmentation.
We show that we achieve similar results compared to an established U-Net segmentation architecture.
The proposed method can be trained in a weakly supervised fashion, as the training samples must be only labeled on image-level.
arXiv Detail & Related papers (2022-02-01T10:26:10Z) - Exploring Cross-Image Pixel Contrast for Semantic Segmentation [130.22216825377618]
We propose a pixel-wise contrastive framework for semantic segmentation in the fully supervised setting.
The core idea is to enforce pixel embeddings belonging to a same semantic class to be more similar than embeddings from different classes.
Our method can be effortlessly incorporated into existing segmentation frameworks without extra overhead during testing.
arXiv Detail & Related papers (2021-01-28T11:35:32Z) - AINet: Association Implantation for Superpixel Segmentation [82.21559299694555]
We propose a novel textbfAssociation textbfImplantation (AI) module to enable the network to explicitly capture the relations between the pixel and its surrounding grids.
Our method could not only achieve state-of-the-art performance but maintain satisfactory inference efficiency.
arXiv Detail & Related papers (2021-01-26T10:40:13Z) - Fully Convolutional Networks for Panoptic Segmentation [91.84686839549488]
We present a conceptually simple, strong, and efficient framework for panoptic segmentation, called Panoptic FCN.
Our approach aims to represent and predict foreground things and background stuff in a unified fully convolutional pipeline.
Panoptic FCN encodes each object instance or stuff category into a specific kernel weight with the proposed kernel generator.
arXiv Detail & Related papers (2020-12-01T18:31:41Z) - Towards Bounding-Box Free Panoptic Segmentation [16.4548904544277]
We introduce a new Bounding-Box Free Network (BBFNet) for panoptic segmentation.
BBFNet predicts coarse watershed levels and uses them to detect large instance candidates where boundaries are well defined.
For smaller instances, whose boundaries are less reliable, BBFNet also predicts instance centers by means of Hough voting followed by mean-shift to reliably detect small objects.
arXiv Detail & Related papers (2020-02-18T16:34:01Z) - The Semantic Mutex Watershed for Efficient Bottom-Up Semantic Instance
Segmentation [15.768804877756384]
We propose a greedy algorithm for joint graph partitioning and labeling.
Due to the algorithm's efficiency it can operate directly on pixels without prior over-segmentation of the image into superpixels.
arXiv Detail & Related papers (2019-12-29T19:48:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.