Revisiting Image Reconstruction for Semi-supervised Semantic
Segmentation
- URL: http://arxiv.org/abs/2303.09794v1
- Date: Fri, 17 Mar 2023 06:31:06 GMT
- Title: Revisiting Image Reconstruction for Semi-supervised Semantic
Segmentation
- Authors: Yuhao Lin, Haiming Xu, Lingqiao Liu, Jinan Zou, Javen Qinfeng Shi
- Abstract summary: We revisit the idea of using image reconstruction as an auxiliary task and incorporate it with a modern semi-supervised semantic segmentation framework.
Surprisingly, we discover that such an old idea in semi-supervised learning can produce results competitive with state-of-the-art semantic segmentation algorithms.
- Score: 16.27277238968567
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Autoencoding, which aims to reconstruct the input images through a bottleneck
latent representation, is one of the classic feature representation learning
strategies. It has been shown effective as an auxiliary task for
semi-supervised learning but has become less popular as more sophisticated
methods have been proposed in recent years. In this paper, we revisit the idea
of using image reconstruction as the auxiliary task and incorporate it with a
modern semi-supervised semantic segmentation framework. Surprisingly, we
discover that such an old idea in semi-supervised learning can produce results
competitive with state-of-the-art semantic segmentation algorithms. By
visualizing the intermediate layer activations of the image reconstruction
module, we show that the feature map channel could correlate well with the
semantic concept, which explains why joint training with the reconstruction
task is helpful for the segmentation task. Motivated by our observation, we
further proposed a modification to the image reconstruction task, aiming to
further disentangle the object clue from the background patterns. From
experiment evaluation on various datasets, we show that using reconstruction as
auxiliary loss can lead to consistent improvements in various datasets and
methods. The proposed method can further lead to significant improvement in
object-centric segmentation tasks.
Related papers
- Towards Task-Compatible Compressible Representations [0.7980273012483663]
We investigate an issue in multi-task learnable compression, in which a representation learned for one task does not positively contribute to the rate-distortion performance of a different task.
In learnable scalable coding, previous work increased the utilization of side-information for input reconstruction by also rewarding input reconstruction when learning this shared representation.
We perform experiments using representations trained for object detection on COCO 2017 and depth estimation on the Cityscapes dataset, and use them to assist in image reconstruction and semantic segmentation tasks.
arXiv Detail & Related papers (2024-05-16T16:47:46Z) - Localized Region Contrast for Enhancing Self-Supervised Learning in
Medical Image Segmentation [27.82940072548603]
We propose a novel contrastive learning framework that integrates Localized Region Contrast (LRC) to enhance existing self-supervised pre-training methods for medical image segmentation.
Our approach involves identifying Super-pixels by Felzenszwalb's algorithm and performing local contrastive learning using a novel contrastive sampling loss.
arXiv Detail & Related papers (2023-04-06T22:43:13Z) - Semantics-Aware Dynamic Localization and Refinement for Referring Image
Segmentation [102.25240608024063]
Referring image segments an image from a language expression.
We develop an algorithm that shifts from being localization-centric to segmentation-language.
Compared to its counterparts, our method is more versatile yet effective.
arXiv Detail & Related papers (2023-03-11T08:42:40Z) - An Empirical Investigation of Representation Learning for Imitation [76.48784376425911]
Recent work in vision, reinforcement learning, and NLP has shown that auxiliary representation learning objectives can reduce the need for large amounts of expensive, task-specific data.
We propose a modular framework for constructing representation learning algorithms, then use our framework to evaluate the utility of representation learning for imitation.
arXiv Detail & Related papers (2022-05-16T11:23:42Z) - Unsupervised Part Discovery from Contrastive Reconstruction [90.88501867321573]
The goal of self-supervised visual representation learning is to learn strong, transferable image representations.
We propose an unsupervised approach to object part discovery and segmentation.
Our method yields semantic parts consistent across fine-grained but visually distinct categories.
arXiv Detail & Related papers (2021-11-11T17:59:42Z) - Leveraging Auxiliary Tasks with Affinity Learning for Weakly Supervised
Semantic Segmentation [88.49669148290306]
We propose a novel weakly supervised multi-task framework called AuxSegNet to leverage saliency detection and multi-label image classification as auxiliary tasks.
Inspired by their similar structured semantics, we also propose to learn a cross-task global pixel-level affinity map from the saliency and segmentation representations.
The learned cross-task affinity can be used to refine saliency predictions and propagate CAM maps to provide improved pseudo labels for both tasks.
arXiv Detail & Related papers (2021-07-25T11:39:58Z) - Self-supervised Segmentation via Background Inpainting [96.10971980098196]
We introduce a self-supervised detection and segmentation approach that can work with single images captured by a potentially moving camera.
We exploit a self-supervised loss function that we exploit to train a proposal-based segmentation network.
We apply our method to human detection and segmentation in images that visually depart from those of standard benchmarks and outperform existing self-supervised methods.
arXiv Detail & Related papers (2020-11-11T08:34:40Z) - Mining Cross-Image Semantics for Weakly Supervised Semantic Segmentation [128.03739769844736]
Two neural co-attentions are incorporated into the classifier to capture cross-image semantic similarities and differences.
In addition to boosting object pattern learning, the co-attention can leverage context from other related images to improve localization map inference.
Our algorithm sets new state-of-the-arts on all these settings, demonstrating well its efficacy and generalizability.
arXiv Detail & Related papers (2020-07-03T21:53:46Z) - Image Embedded Segmentation: Uniting Supervised and Unsupervised
Objectives for Segmenting Histopathological Images [0.0]
This paper presents a new regularization method to train a fully convolutional network for semantic tissue segmentation.
It relies on the benefit of unsupervised learning, in the form of image reconstruction, for network training.
Our experiments demonstrate that it leads to better segmentation results in these datasets, compared to its counterparts.
arXiv Detail & Related papers (2020-01-30T08:09:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.