Self Correspondence Distillation for End-to-End Weakly-Supervised
Semantic Segmentation
- URL: http://arxiv.org/abs/2302.13765v1
- Date: Mon, 27 Feb 2023 13:46:40 GMT
- Title: Self Correspondence Distillation for End-to-End Weakly-Supervised
Semantic Segmentation
- Authors: Rongtao Xu, Changwei Wang, Jiaxi Sun, Shibiao Xu, Weiliang Meng,
Xiaopeng Zhang
- Abstract summary: We propose a novel Self Correspondence Distillation (SCD) method to refine pseudo-labels without introducing external supervision.
In addition, we design a Variation-aware Refine Module to enhance the local consistency of pseudo-labels.
Our method significantly outperforms other state-of-the-art methods.
- Score: 13.623713806739271
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Efficiently training accurate deep models for weakly supervised semantic
segmentation (WSSS) with image-level labels is challenging and important.
Recently, end-to-end WSSS methods have become the focus of research due to
their high training efficiency. However, current methods suffer from
insufficient extraction of comprehensive semantic information, resulting in
low-quality pseudo-labels and sub-optimal solutions for end-to-end WSSS. To
this end, we propose a simple and novel Self Correspondence Distillation (SCD)
method to refine pseudo-labels without introducing external supervision. Our
SCD enables the network to utilize feature correspondence derived from itself
as a distillation target, which can enhance the network's feature learning
process by complementing semantic information. In addition, to further improve
the segmentation accuracy, we design a Variation-aware Refine Module to enhance
the local consistency of pseudo-labels by computing pixel-level variation.
Finally, we present an efficient end-to-end Transformer-based framework (TSCD)
via SCD and Variation-aware Refine Module for the accurate WSSS task. Extensive
experiments on the PASCAL VOC 2012 and MS COCO 2014 datasets demonstrate that
our method significantly outperforms other state-of-the-art methods.
Our code is available at
{https://github.com/Rongtao-Xu/RepresentationLearning/tree/main/SCD-AAAI2023}.
Related papers
- AdaSemiCD: An Adaptive Semi-Supervised Change Detection Method Based on Pseudo-Label Evaluation [0.0]
We present an adaptive dynamic semi-supervised learning method, AdaCD, to improve the use of pseudo-labels and optimize the training process.
Experimental results from LEVIR-CD, WHU-CD, and CDD datasets validate the efficacy and universality of our proposed adaptive training framework.
arXiv Detail & Related papers (2024-11-12T12:35:34Z) - Unsupervised Domain Adaptation for Semantic Segmentation with Pseudo
Label Self-Refinement [9.69089112870202]
We propose an auxiliary pseudo-label refinement network (PRN) for online refining of the pseudo labels and also localizing the pixels whose predicted labels are likely to be noisy.
We evaluate our approach on benchmark datasets with three different domain shifts, and our approach consistently performs significantly better than the previous state-of-the-art methods.
arXiv Detail & Related papers (2023-10-25T20:31:07Z) - Hard-aware Instance Adaptive Self-training for Unsupervised Cross-domain
Semantic Segmentation [18.807921765977415]
We propose a hard-aware instance adaptive self-training framework for UDA on the task of semantic segmentation.
We develop a novel pseudo-label generation strategy with an instance adaptive selector.
Experiments on GTA5 to Cityscapes, SYNTHIA to Cityscapes, and Cityscapes to Oxford RobotCar demonstrate the superior performance of our approach.
arXiv Detail & Related papers (2023-02-14T11:52:26Z) - Dense FixMatch: a simple semi-supervised learning method for pixel-wise
prediction tasks [68.36996813591425]
We propose Dense FixMatch, a simple method for online semi-supervised learning of dense and structured prediction tasks.
We enable the application of FixMatch in semi-supervised learning problems beyond image classification by adding a matching operation on the pseudo-labels.
Dense FixMatch significantly improves results compared to supervised learning using only labeled data, approaching its performance with 1/4 of the labeled samples.
arXiv Detail & Related papers (2022-10-18T15:02:51Z) - Dense Learning based Semi-Supervised Object Detection [46.885301243656045]
Semi-supervised object detection (SSOD) aims to facilitate the training and deployment of object detectors with the help of a large amount of unlabeled data.
In this paper, we propose a DenSe Learning based anchor-free SSOD algorithm.
Experiments are conducted on MS-COCO and PASCAL-VOC, and the results show that our proposed DSL method records new state-of-the-art SSOD performance.
arXiv Detail & Related papers (2022-04-15T02:31:02Z) - Learning Affinity from Attention: End-to-End Weakly-Supervised Semantic
Segmentation with Transformers [44.757309147148035]
We introduce Transformers, which naturally integrate global information, to generate more integral initial pseudo labels for end-to-end WSSS.
Motivated by the inherent consistency between the self-attention in Transformers and the semantic affinity, we propose an Affinity from Attention (AFA) module.
We also devise a Pixel-Adaptive Refinement module that incorporates low-level image appearance information to refine the pseudo labels.
arXiv Detail & Related papers (2022-03-05T06:07:17Z) - Real-Time Scene Text Detection with Differentiable Binarization and
Adaptive Scale Fusion [62.269219152425556]
segmentation-based scene text detection methods have drawn extensive attention in the scene text detection field.
We propose a Differentiable Binarization (DB) module that integrates the binarization process into a segmentation network.
An efficient Adaptive Scale Fusion (ASF) module is proposed to improve the scale robustness by fusing features of different scales adaptively.
arXiv Detail & Related papers (2022-02-21T15:30:14Z) - Improving Semi-Supervised and Domain-Adaptive Semantic Segmentation with
Self-Supervised Depth Estimation [94.16816278191477]
We present a framework for semi-adaptive and domain-supervised semantic segmentation.
It is enhanced by self-supervised monocular depth estimation trained only on unlabeled image sequences.
We validate the proposed model on the Cityscapes dataset.
arXiv Detail & Related papers (2021-08-28T01:33:38Z) - A Simple Baseline for Semi-supervised Semantic Segmentation with Strong
Data Augmentation [74.8791451327354]
We propose a simple yet effective semi-supervised learning framework for semantic segmentation.
A set of simple design and training techniques can collectively improve the performance of semi-supervised semantic segmentation significantly.
Our method achieves state-of-the-art results in the semi-supervised settings on the Cityscapes and Pascal VOC datasets.
arXiv Detail & Related papers (2021-04-15T06:01:39Z) - PseudoSeg: Designing Pseudo Labels for Semantic Segmentation [78.35515004654553]
We present a re-design of pseudo-labeling to generate structured pseudo labels for training with unlabeled or weakly-labeled data.
We demonstrate the effectiveness of the proposed pseudo-labeling strategy in both low-data and high-data regimes.
arXiv Detail & Related papers (2020-10-19T17:59:30Z) - Deep Semi-supervised Knowledge Distillation for Overlapping Cervical
Cell Instance Segmentation [54.49894381464853]
We propose to leverage both labeled and unlabeled data for instance segmentation with improved accuracy by knowledge distillation.
We propose a novel Mask-guided Mean Teacher framework with Perturbation-sensitive Sample Mining.
Experiments show that the proposed method improves the performance significantly compared with the supervised method learned from labeled data only.
arXiv Detail & Related papers (2020-07-21T13:27:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.