Transfer beyond the Field of View: Dense Panoramic Semantic Segmentation
via Unsupervised Domain Adaptation
- URL: http://arxiv.org/abs/2110.11062v1
- Date: Thu, 21 Oct 2021 11:22:05 GMT
- Title: Transfer beyond the Field of View: Dense Panoramic Semantic Segmentation
via Unsupervised Domain Adaptation
- Authors: Jiaming Zhang, Chaoxiang Ma, Kailun Yang, Alina Roitberg, Kunyu Peng,
Rainer Stiefelhagen
- Abstract summary: We formalize the task of unsupervised domain adaptation for panoramic semantic segmentation.
DensePASS is a novel dataset for panoramic segmentation under cross-domain conditions.
We introduce P2PDA - a generic framework for Pinhole-to-Panoramic semantic segmentation.
- Score: 30.104947024614127
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Autonomous vehicles clearly benefit from the expanded Field of View (FoV) of
360-degree sensors, but modern semantic segmentation approaches rely heavily on
annotated training data which is rarely available for panoramic images. We look
at this problem from the perspective of domain adaptation and bring panoramic
semantic segmentation to a setting, where labelled training data originates
from a different distribution of conventional pinhole camera images. To achieve
this, we formalize the task of unsupervised domain adaptation for panoramic
semantic segmentation and collect DensePASS - a novel densely annotated dataset
for panoramic segmentation under cross-domain conditions, specifically built to
study the Pinhole-to-Panoramic domain shift and accompanied with pinhole camera
training examples obtained from Cityscapes. DensePASS covers both, labelled-
and unlabelled 360-degree images, with the labelled data comprising 19 classes
which explicitly fit the categories available in the source (i.e. pinhole)
domain. Since data-driven models are especially susceptible to changes in data
distribution, we introduce P2PDA - a generic framework for Pinhole-to-Panoramic
semantic segmentation which addresses the challenge of domain divergence with
different variants of attention-augmented domain adaptation modules, enabling
the transfer in output-, feature-, and feature confidence spaces. P2PDA
intertwines uncertainty-aware adaptation using confidence values regulated
on-the-fly through attention heads with discrepant predictions. Our framework
facilitates context exchange when learning domain correspondences and
dramatically improves the adaptation performance of accuracy- and
efficiency-focused models. Comprehensive experiments verify that our framework
clearly surpasses unsupervised domain adaptation- and specialized panoramic
segmentation approaches.
Related papers
- PiPa++: Towards Unification of Domain Adaptive Semantic Segmentation via Self-supervised Learning [34.786268652516355]
Unsupervised domain adaptive segmentation aims to improve the segmentation accuracy of models on target domains without relying on labeled data from those domains.
It seeks to align the feature representations of the source domain (where labeled data is available) and the target domain (where only unlabeled data is present)
arXiv Detail & Related papers (2024-07-24T08:53:29Z) - Exploring Consistency in Cross-Domain Transformer for Domain Adaptive
Semantic Segmentation [51.10389829070684]
Domain gap can cause discrepancies in self-attention.
Due to this gap, the transformer attends to spurious regions or pixels, which deteriorates accuracy on the target domain.
We propose adaptation on attention maps with cross-domain attention layers.
arXiv Detail & Related papers (2022-11-27T02:40:33Z) - PiPa: Pixel- and Patch-wise Self-supervised Learning for Domain
Adaptative Semantic Segmentation [100.6343963798169]
Unsupervised Domain Adaptation (UDA) aims to enhance the generalization of the learned model to other domains.
We propose a unified pixel- and patch-wise self-supervised learning framework, called PiPa, for domain adaptive semantic segmentation.
arXiv Detail & Related papers (2022-11-14T18:31:24Z) - Panoramic Panoptic Segmentation: Insights Into Surrounding Parsing for
Mobile Agents via Unsupervised Contrastive Learning [93.6645991946674]
We introduce panoramic panoptic segmentation, as the most holistic scene understanding.
A complete surrounding understanding provides a maximum of information to a mobile agent.
We propose a framework which allows model training on standard pinhole images and transfers the learned features to a different domain.
arXiv Detail & Related papers (2022-06-21T20:07:15Z) - DensePASS: Dense Panoramic Semantic Segmentation via Unsupervised Domain
Adaptation with Attention-Augmented Context Exchange [32.29797061415896]
We formalize the task of unsupervised domain adaptation for panoramic semantic segmentation.
A network trained on labelled examples from the source domain of pinhole camera data is deployed in a different target domain of panoramic images.
We build a generic framework for cross-domain panoramic semantic segmentation based on different variants of attention-augmented domain adaptation modules.
arXiv Detail & Related papers (2021-08-13T20:15:46Z) - Domain Adaptation for Semantic Segmentation via Patch-Wise Contrastive
Learning [62.7588467386166]
We leverage contrastive learning to bridge the domain gap by aligning the features of structurally similar label patches across domains.
Our approach consistently outperforms state-of-the-art unsupervised and semi-supervised methods on two challenging domain adaptive segmentation tasks.
arXiv Detail & Related papers (2021-04-22T13:39:12Z) - Panoramic Panoptic Segmentation: Towards Complete Surrounding
Understanding via Unsupervised Contrastive Learning [97.37544023666833]
We introduce panoramic panoptic segmentation as the most holistic scene understanding.
A complete surrounding understanding provides a maximum of information to the agent.
We propose a framework which allows model training on standard pinhole images and transfers the learned features to a different domain.
arXiv Detail & Related papers (2021-03-01T09:37:27Z) - Domain Adaptation on Semantic Segmentation for Aerial Images [3.946367634483361]
We propose a novel unsupervised domain adaptation framework to address domain shift in semantic image segmentation.
We also apply entropy minimization on the target domain to produce high-confident prediction.
We show improvement over state-of-the-art methods in terms of various metrics.
arXiv Detail & Related papers (2020-12-03T20:58:27Z) - Phase Consistent Ecological Domain Adaptation [76.75730500201536]
We focus on the task of semantic segmentation, where annotated synthetic data are aplenty, but annotating real data is laborious.
The first criterion, inspired by visual psychophysics, is that the map between the two image domains be phase-preserving.
The second criterion aims to leverage ecological statistics, or regularities in the scene which are manifest in any image of it, regardless of the characteristics of the illuminant or the imaging sensor.
arXiv Detail & Related papers (2020-04-10T06:58:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.