Panoramic Panoptic Segmentation: Insights Into Surrounding Parsing for
Mobile Agents via Unsupervised Contrastive Learning
- URL: http://arxiv.org/abs/2206.10711v1
- Date: Tue, 21 Jun 2022 20:07:15 GMT
- Title: Panoramic Panoptic Segmentation: Insights Into Surrounding Parsing for
Mobile Agents via Unsupervised Contrastive Learning
- Authors: Alexander Jaus, Kailun Yang, Rainer Stiefelhagen
- Abstract summary: We introduce panoramic panoptic segmentation, as the most holistic scene understanding.
A complete surrounding understanding provides a maximum of information to a mobile agent.
We propose a framework which allows model training on standard pinhole images and transfers the learned features to a different domain.
- Score: 93.6645991946674
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this work, we introduce panoramic panoptic segmentation, as the most
holistic scene understanding, both in terms of Field of View (FoV) and
image-level understanding for standard camera-based input. A complete
surrounding understanding provides a maximum of information to a mobile agent,
which is essential for any intelligent vehicle in order to make informed
decisions in a safety-critical dynamic environment such as real-world traffic.
In order to overcome the lack of annotated panoramic images, we propose a
framework which allows model training on standard pinhole images and transfers
the learned features to a different domain in a cost-minimizing way. Using our
proposed method with dense contrastive learning, we manage to achieve
significant improvements over a non-adapted approach. Depending on the
efficient panoptic segmentation architecture, we can improve 3.5-6.5% measured
in Panoptic Quality (PQ) over non-adapted models on our established Wild
Panoramic Panoptic Segmentation (WildPPS) dataset. Furthermore, our efficient
framework does not need access to the images of the target domain, making it a
feasible domain generalization approach suitable for a limited hardware
setting. As additional contributions, we publish WildPPS: The first panoramic
panoptic image dataset to foster progress in surrounding perception and explore
a novel training procedure combining supervised and contrastive training.
Related papers
- Few-Shot Panoptic Segmentation With Foundation Models [23.231014713335664]
We propose to leverage task-agnostic image features to enable few-shot panoptic segmentation by presenting Segmenting Panoptic Information with Nearly 0 labels (SPINO)
In detail, our method combines a DINOv2 backbone with lightweight network heads for semantic segmentation and boundary estimation.
We show that our approach, albeit being trained with only ten annotated images, predicts high-quality pseudo-labels that can be used with any existing panoptic segmentation method.
arXiv Detail & Related papers (2023-09-19T16:09:01Z) - Rethinking Range View Representation for LiDAR Segmentation [66.73116059734788]
"Many-to-one" mapping, semantic incoherence, and shape deformation are possible impediments against effective learning from range view projections.
We present RangeFormer, a full-cycle framework comprising novel designs across network architecture, data augmentation, and post-processing.
We show that, for the first time, a range view method is able to surpass the point, voxel, and multi-view fusion counterparts in the competing LiDAR semantic and panoptic segmentation benchmarks.
arXiv Detail & Related papers (2023-03-09T16:13:27Z) - Transfer beyond the Field of View: Dense Panoramic Semantic Segmentation
via Unsupervised Domain Adaptation [30.104947024614127]
We formalize the task of unsupervised domain adaptation for panoramic semantic segmentation.
DensePASS is a novel dataset for panoramic segmentation under cross-domain conditions.
We introduce P2PDA - a generic framework for Pinhole-to-Panoramic semantic segmentation.
arXiv Detail & Related papers (2021-10-21T11:22:05Z) - Navigation-Oriented Scene Understanding for Robotic Autonomy: Learning
to Segment Driveability in Egocentric Images [25.350677396144075]
This work tackles scene understanding for outdoor robotic navigation, solely relying on images captured by an on-board camera.
We segment egocentric images directly in terms of how a robot can navigate in them, and tailor the learning problem to an autonomous navigation task.
We present a generic and scalable affordance-based definition consisting of 3 driveability levels which can be applied to arbitrary scenes.
arXiv Detail & Related papers (2021-09-15T12:25:56Z) - DeepPanoContext: Panoramic 3D Scene Understanding with Holistic Scene
Context Graph and Relation-based Optimization [66.25948693095604]
We propose a novel method for panoramic 3D scene understanding which recovers the 3D room layout and the shape, pose, position, and semantic category for each object from a single full-view panorama image.
Experiments demonstrate that our method outperforms existing methods on panoramic scene understanding in terms of both geometry accuracy and object arrangement.
arXiv Detail & Related papers (2021-08-24T13:55:29Z) - DensePASS: Dense Panoramic Semantic Segmentation via Unsupervised Domain
Adaptation with Attention-Augmented Context Exchange [32.29797061415896]
We formalize the task of unsupervised domain adaptation for panoramic semantic segmentation.
A network trained on labelled examples from the source domain of pinhole camera data is deployed in a different target domain of panoramic images.
We build a generic framework for cross-domain panoramic semantic segmentation based on different variants of attention-augmented domain adaptation modules.
arXiv Detail & Related papers (2021-08-13T20:15:46Z) - Panoramic Panoptic Segmentation: Towards Complete Surrounding
Understanding via Unsupervised Contrastive Learning [97.37544023666833]
We introduce panoramic panoptic segmentation as the most holistic scene understanding.
A complete surrounding understanding provides a maximum of information to the agent.
We propose a framework which allows model training on standard pinhole images and transfers the learned features to a different domain.
arXiv Detail & Related papers (2021-03-01T09:37:27Z) - Self-supervised Human Detection and Segmentation via Multi-view
Consensus [116.92405645348185]
We propose a multi-camera framework in which geometric constraints are embedded in the form of multi-view consistency during training.
We show that our approach outperforms state-of-the-art self-supervised person detection and segmentation techniques on images that visually depart from those of standard benchmarks.
arXiv Detail & Related papers (2020-12-09T15:47:21Z) - Self-supervised Equivariant Attention Mechanism for Weakly Supervised
Semantic Segmentation [93.83369981759996]
We propose a self-supervised equivariant attention mechanism (SEAM) to discover additional supervision and narrow the gap.
Our method is based on the observation that equivariance is an implicit constraint in fully supervised semantic segmentation.
We propose consistency regularization on predicted CAMs from various transformed images to provide self-supervision for network learning.
arXiv Detail & Related papers (2020-04-09T14:57:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.