Context Prior for Scene Segmentation
- URL: http://arxiv.org/abs/2004.01547v1
- Date: Fri, 3 Apr 2020 13:16:32 GMT
- Title: Context Prior for Scene Segmentation
- Authors: Changqian Yu, Jingbo Wang, Changxin Gao, Gang Yu, Chunhua Shen, Nong
Sang
- Abstract summary: We develop a Context Prior with the supervision of the Affinity Loss.
The learned Context Prior extracts the pixels belonging to the same category, while the reversed prior focuses on the pixels of different classes.
Our algorithm achieves 46.3% mIoU on ADE20K, 53.9% mIoU on PASCAL-Context, and 81.3% mIoU on Cityscapes.
- Score: 118.46210049742993
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recent works have widely explored the contextual dependencies to achieve more
accurate segmentation results. However, most approaches rarely distinguish
different types of contextual dependencies, which may pollute the scene
understanding. In this work, we directly supervise the feature aggregation to
distinguish the intra-class and inter-class context clearly. Specifically, we
develop a Context Prior with the supervision of the Affinity Loss. Given an
input image and corresponding ground truth, Affinity Loss constructs an ideal
affinity map to supervise the learning of Context Prior. The learned Context
Prior extracts the pixels belonging to the same category, while the reversed
prior focuses on the pixels of different classes. Embedded into a conventional
deep CNN, the proposed Context Prior Layer can selectively capture the
intra-class and inter-class contextual dependencies, leading to robust feature
representation. To validate the effectiveness, we design an effective Context
Prior Network (CPNet). Extensive quantitative and qualitative evaluations
demonstrate that the proposed model performs favorably against state-of-the-art
semantic segmentation approaches. More specifically, our algorithm achieves
46.3% mIoU on ADE20K, 53.9% mIoU on PASCAL-Context, and 81.3% mIoU on
Cityscapes. Code is available at https://git.io/ContextPrior.
Related papers
- Copy-Pasting Coherent Depth Regions Improves Contrastive Learning for
Urban-Scene Segmentation [14.742591401997048]
We group coherent, semantically related pixels into coherent depth regions given their estimated depth.
Our method surpasses the previous state-of-the-art baseline by +7.14% in mIoU on Cityscapes and +6.65% on KITTI.
arXiv Detail & Related papers (2022-11-25T12:52:08Z) - CoDo: Contrastive Learning with Downstream Background Invariance for
Detection [10.608660802917214]
We propose a novel object-level self-supervised learning method, called Contrastive learning with Downstream background invariance (CoDo)
The pretext task is converted to focus on instance location modeling for various backgrounds, especially for downstream datasets.
Experiments on MSCOCO demonstrate that the proposed CoDo with common backbones, ResNet50-FPN, yields strong transfer learning results for object detection.
arXiv Detail & Related papers (2022-05-10T01:26:15Z) - Unsupervised Part Discovery from Contrastive Reconstruction [90.88501867321573]
The goal of self-supervised visual representation learning is to learn strong, transferable image representations.
We propose an unsupervised approach to object part discovery and segmentation.
Our method yields semantic parts consistent across fine-grained but visually distinct categories.
arXiv Detail & Related papers (2021-11-11T17:59:42Z) - Revisiting Contrastive Methods for Unsupervised Learning of Visual
Representations [78.12377360145078]
Contrastive self-supervised learning has outperformed supervised pretraining on many downstream tasks like segmentation and object detection.
In this paper, we first study how biases in the dataset affect existing methods.
We show that current contrastive approaches work surprisingly well across: (i) object- versus scene-centric, (ii) uniform versus long-tailed and (iii) general versus domain-specific datasets.
arXiv Detail & Related papers (2021-06-10T17:59:13Z) - Exploring Cross-Image Pixel Contrast for Semantic Segmentation [130.22216825377618]
We propose a pixel-wise contrastive framework for semantic segmentation in the fully supervised setting.
The core idea is to enforce pixel embeddings belonging to a same semantic class to be more similar than embeddings from different classes.
Our method can be effortlessly incorporated into existing segmentation frameworks without extra overhead during testing.
arXiv Detail & Related papers (2021-01-28T11:35:32Z) - Dense Contrastive Learning for Self-Supervised Visual Pre-Training [102.15325936477362]
We present dense contrastive learning, which implements self-supervised learning by optimizing a pairwise contrastive (dis)similarity loss at the pixel level between two views of input images.
Compared to the baseline method MoCo-v2, our method introduces negligible computation overhead (only 1% slower)
arXiv Detail & Related papers (2020-11-18T08:42:32Z) - Contextual Diversity for Active Learning [9.546771465714876]
Large datasets restrict the use of deep convolutional neural networks (CNNs) for many practical applications.
We introduce the notion of contextual diversity that captures the confusion associated with spatially co-occurring classes.
Our studies show clear advantages of using contextual diversity for active learning.
arXiv Detail & Related papers (2020-08-13T07:04:15Z) - SCAN: Learning to Classify Images without Labels [73.69513783788622]
We advocate a two-step approach where feature learning and clustering are decoupled.
A self-supervised task from representation learning is employed to obtain semantically meaningful features.
We obtain promising results on ImageNet, and outperform several semi-supervised learning methods in the low-data regime.
arXiv Detail & Related papers (2020-05-25T18:12:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.