CPRAL: Collaborative Panoptic-Regional Active Learning for Semantic
Segmentation
- URL: http://arxiv.org/abs/2112.05975v1
- Date: Sat, 11 Dec 2021 13:13:13 GMT
- Title: CPRAL: Collaborative Panoptic-Regional Active Learning for Semantic
Segmentation
- Authors: Yu Qiao, Jincheng Zhu, Chengjiang Long, Zeyao Zhang, Yuxin Wang,
Zhenjun Du, Xin Yang
- Abstract summary: We propose a Collaborative Panoptic-Regional Active Learning framework (CPRAL) to address the semantic segmentation task.
Considering the class imbalance in the segmentation dataset, we import a Regional Gaussian Attention module (RGA) to achieve semantics-biased selection.
We show that CPRAL outperforms the cutting-edge methods with impressive results and less labeling proportion.
- Score: 35.11139361684248
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Acquiring the most representative examples via active learning (AL) can
benefit many data-dependent computer vision tasks by minimizing efforts of
image-level or pixel-wise annotations. In this paper, we propose a novel
Collaborative Panoptic-Regional Active Learning framework (CPRAL) to address
the semantic segmentation task. For a small batch of images initially sampled
with pixel-wise annotations, we employ panoptic information to initially select
unlabeled samples. Considering the class imbalance in the segmentation dataset,
we import a Regional Gaussian Attention module (RGA) to achieve
semantics-biased selection. The subset is highlighted by vote entropy and then
attended by Gaussian kernels to maximize the biased regions. We also propose a
Contextual Labels Extension (CLE) to boost regional annotations with contextual
attention guidance. With the collaboration of semantics-agnostic panoptic
matching and regionbiased selection and extension, our CPRAL can strike a
balance between labeling efforts and performance and compromise the semantics
distribution. We perform extensive experiments on Cityscapes and BDD10K
datasets and show that CPRAL outperforms the cutting-edge methods with
impressive results and less labeling proportion.
Related papers
- Auxiliary Tasks Enhanced Dual-affinity Learning for Weakly Supervised
Semantic Segmentation [79.05949524349005]
We propose AuxSegNet+, a weakly supervised auxiliary learning framework to explore the rich information from saliency maps.
We also propose a cross-task affinity learning mechanism to learn pixel-level affinities from the saliency and segmentation feature maps.
arXiv Detail & Related papers (2024-03-02T10:03:21Z) - Exploring Open-Vocabulary Semantic Segmentation without Human Labels [76.15862573035565]
We present ZeroSeg, a novel method that leverages the existing pretrained vision-language model (VL) to train semantic segmentation models.
ZeroSeg overcomes this by distilling the visual concepts learned by VL models into a set of segment tokens, each summarizing a localized region of the target image.
Our approach achieves state-of-the-art performance when compared to other zero-shot segmentation methods under the same training data.
arXiv Detail & Related papers (2023-06-01T08:47:06Z) - Localized Region Contrast for Enhancing Self-Supervised Learning in
Medical Image Segmentation [27.82940072548603]
We propose a novel contrastive learning framework that integrates Localized Region Contrast (LRC) to enhance existing self-supervised pre-training methods for medical image segmentation.
Our approach involves identifying Super-pixels by Felzenszwalb's algorithm and performing local contrastive learning using a novel contrastive sampling loss.
arXiv Detail & Related papers (2023-04-06T22:43:13Z) - Context Label Learning: Improving Background Class Representations in
Semantic Segmentation [23.79946807540805]
We find that neural networks trained with heterogeneous background struggle to map the corresponding contextual samples to compact clusters in feature space.
We propose context label learning (CoLab) to improve the context representations by decomposing the background class into several subclasses.
The results demonstrate that CoLab can guide the segmentation model to map the logits of background samples away from the decision boundary.
arXiv Detail & Related papers (2022-12-16T11:52:15Z) - Improving Semi-Supervised and Domain-Adaptive Semantic Segmentation with
Self-Supervised Depth Estimation [94.16816278191477]
We present a framework for semi-adaptive and domain-supervised semantic segmentation.
It is enhanced by self-supervised monocular depth estimation trained only on unlabeled image sequences.
We validate the proposed model on the Cityscapes dataset.
arXiv Detail & Related papers (2021-08-28T01:33:38Z) - Leveraging Auxiliary Tasks with Affinity Learning for Weakly Supervised
Semantic Segmentation [88.49669148290306]
We propose a novel weakly supervised multi-task framework called AuxSegNet to leverage saliency detection and multi-label image classification as auxiliary tasks.
Inspired by their similar structured semantics, we also propose to learn a cross-task global pixel-level affinity map from the saliency and segmentation representations.
The learned cross-task affinity can be used to refine saliency predictions and propagate CAM maps to provide improved pseudo labels for both tasks.
arXiv Detail & Related papers (2021-07-25T11:39:58Z) - Semantic Segmentation with Generative Models: Semi-Supervised Learning
and Strong Out-of-Domain Generalization [112.68171734288237]
We propose a novel framework for discriminative pixel-level tasks using a generative model of both images and labels.
We learn a generative adversarial network that captures the joint image-label distribution and is trained efficiently using a large set of unlabeled images.
We demonstrate strong in-domain performance compared to several baselines, and are the first to showcase extreme out-of-domain generalization.
arXiv Detail & Related papers (2021-04-12T21:41:25Z) - Deep Active Learning for Joint Classification & Segmentation with Weak
Annotator [22.271760669551817]
CNN visualization and interpretation methods, like class-activation maps (CAMs), are typically used to highlight the image regions linked to class predictions.
We propose an active learning framework, which progressively integrates pixel-level annotations during training.
Our results indicate that, by simply using random sample selection, the proposed approach can significantly outperform state-of-the-art CAMs and AL methods.
arXiv Detail & Related papers (2020-10-10T03:25:54Z) - Manifold-driven Attention Maps for Weakly Supervised Segmentation [9.289524646688244]
We propose a manifold driven attention-based network to enhance visual salient regions.
Our method generates superior attention maps directly during inference without the need of extra computations.
arXiv Detail & Related papers (2020-04-07T00:03:28Z) - Weakly-Supervised Semantic Segmentation by Iterative Affinity Learning [86.45526827323954]
Weakly-supervised semantic segmentation is a challenging task as no pixel-wise label information is provided for training.
We propose an iterative algorithm to learn such pairwise relations.
We show that the proposed algorithm performs favorably against the state-of-the-art methods.
arXiv Detail & Related papers (2020-02-19T10:32:03Z) - Reinforced active learning for image segmentation [34.096237671643145]
We present a new active learning strategy for semantic segmentation based on deep reinforcement learning (RL)
An agent learns a policy to select a subset of small informative image regions -- opposed to entire images -- to be labeled from a pool of unlabeled data.
Our method proposes a new modification of the deep Q-network (DQN) formulation for active learning, adapting it to the large-scale nature of semantic segmentation problems.
arXiv Detail & Related papers (2020-02-16T14:03:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.