Leveraging Auxiliary Tasks with Affinity Learning for Weakly Supervised
Semantic Segmentation
- URL: http://arxiv.org/abs/2107.11787v2
- Date: Tue, 27 Jul 2021 02:15:27 GMT
- Title: Leveraging Auxiliary Tasks with Affinity Learning for Weakly Supervised
Semantic Segmentation
- Authors: Lian Xu, Wanli Ouyang, Mohammed Bennamoun, Farid Boussaid, Ferdous
Sohel, Dan Xu
- Abstract summary: We propose a novel weakly supervised multi-task framework called AuxSegNet to leverage saliency detection and multi-label image classification as auxiliary tasks.
Inspired by their similar structured semantics, we also propose to learn a cross-task global pixel-level affinity map from the saliency and segmentation representations.
The learned cross-task affinity can be used to refine saliency predictions and propagate CAM maps to provide improved pseudo labels for both tasks.
- Score: 88.49669148290306
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Semantic segmentation is a challenging task in the absence of densely
labelled data. Only relying on class activation maps (CAM) with image-level
labels provides deficient segmentation supervision. Prior works thus consider
pre-trained models to produce coarse saliency maps to guide the generation of
pseudo segmentation labels. However, the commonly used off-line heuristic
generation process cannot fully exploit the benefits of these coarse saliency
maps. Motivated by the significant inter-task correlation, we propose a novel
weakly supervised multi-task framework termed as AuxSegNet, to leverage
saliency detection and multi-label image classification as auxiliary tasks to
improve the primary task of semantic segmentation using only image-level
ground-truth labels. Inspired by their similar structured semantics, we also
propose to learn a cross-task global pixel-level affinity map from the saliency
and segmentation representations. The learned cross-task affinity can be used
to refine saliency predictions and propagate CAM maps to provide improved
pseudo labels for both tasks. The mutual boost between pseudo label updating
and cross-task affinity learning enables iterative improvements on segmentation
performance. Extensive experiments demonstrate the effectiveness of the
proposed auxiliary learning network structure and the cross-task affinity
learning method. The proposed approach achieves state-of-the-art weakly
supervised segmentation performance on the challenging PASCAL VOC 2012 and MS
COCO benchmarks.
Related papers
- iSeg: An Iterative Refinement-based Framework for Training-free Segmentation [85.58324416386375]
We present a deep experimental analysis on iteratively refining cross-attention map with self-attention map.
We propose an effective iterative refinement framework for training-free segmentation, named iSeg.
Our proposed iSeg achieves an absolute gain of 3.8% in terms of mIoU compared to the best existing training-free approach in literature.
arXiv Detail & Related papers (2024-09-05T03:07:26Z) - Auxiliary Tasks Enhanced Dual-affinity Learning for Weakly Supervised
Semantic Segmentation [79.05949524349005]
We propose AuxSegNet+, a weakly supervised auxiliary learning framework to explore the rich information from saliency maps.
We also propose a cross-task affinity learning mechanism to learn pixel-level affinities from the saliency and segmentation feature maps.
arXiv Detail & Related papers (2024-03-02T10:03:21Z) - Semantic Connectivity-Driven Pseudo-labeling for Cross-domain
Segmentation [89.41179071022121]
Self-training is a prevailing approach in cross-domain semantic segmentation.
We propose a novel approach called Semantic Connectivity-driven pseudo-labeling.
This approach formulates pseudo-labels at the connectivity level and thus can facilitate learning structured and low-noise semantics.
arXiv Detail & Related papers (2023-12-11T12:29:51Z) - Label-efficient Segmentation via Affinity Propagation [27.016747627689288]
Weakly-supervised segmentation with label-efficient sparse annotations has attracted increasing research attention to reduce the cost of laborious pixel-wise labeling process.
We formulate the affinity modeling as an affinity propagation process, and propose a local and a global pairwise affinity terms to generate accurate soft pseudo labels.
An efficient algorithm is also developed to reduce significantly the computational cost.
arXiv Detail & Related papers (2023-10-16T15:54:09Z) - Semantic Contrastive Bootstrapping for Single-positive Multi-label
Recognition [36.3636416735057]
We present a semantic contrastive bootstrapping (Scob) approach to gradually recover the cross-object relationships.
We then propose a recurrent semantic masked transformer to extract iconic object-level representations.
Extensive experimental results demonstrate that the proposed joint learning framework surpasses the state-of-the-art models.
arXiv Detail & Related papers (2023-07-15T01:59:53Z) - Learning to Relate Depth and Semantics for Unsupervised Domain
Adaptation [87.1188556802942]
We present an approach for encoding visual task relationships to improve model performance in an Unsupervised Domain Adaptation (UDA) setting.
We propose a novel Cross-Task Relation Layer (CTRL), which encodes task dependencies between the semantic and depth predictions.
Furthermore, we propose an Iterative Self-Learning (ISL) training scheme, which exploits semantic pseudo-labels to provide extra supervision on the target domain.
arXiv Detail & Related papers (2021-05-17T13:42:09Z) - Weakly-Supervised Semantic Segmentation by Iterative Affinity Learning [86.45526827323954]
Weakly-supervised semantic segmentation is a challenging task as no pixel-wise label information is provided for training.
We propose an iterative algorithm to learn such pairwise relations.
We show that the proposed algorithm performs favorably against the state-of-the-art methods.
arXiv Detail & Related papers (2020-02-19T10:32:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.