SceneAdapt: Scene-based domain adaptation for semantic segmentation
using adversarial learning
- URL: http://arxiv.org/abs/2006.10386v1
- Date: Thu, 18 Jun 2020 09:43:31 GMT
- Title: SceneAdapt: Scene-based domain adaptation for semantic segmentation
using adversarial learning
- Authors: Daniele Di Mauro, Antonino Furnari, Giuseppe Patan\`e, Sebastiano
Battiato, Giovanni Maria Farinella
- Abstract summary: SceneAdapt is a method for scene adaptation of semantic segmentation algorithms based on adversarial learning.
To encourage research on this topic, we made our code available at our web page: https://iplab.dmi.unict.it/ParkSmartSceneAdaptation/.
- Score: 24.37746089471516
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Semantic segmentation methods have achieved outstanding performance thanks to
deep learning. Nevertheless, when such algorithms are deployed to new contexts
not seen during training, it is necessary to collect and label scene-specific
data in order to adapt them to the new domain using fine-tuning. This process
is required whenever an already installed camera is moved or a new camera is
introduced in a camera network due to the different scene layouts induced by
the different viewpoints. To limit the amount of additional training data to be
collected, it would be ideal to train a semantic segmentation method using
labeled data already available and only unlabeled data coming from the new
camera. We formalize this problem as a domain adaptation task and introduce a
novel dataset of urban scenes with the related semantic labels. As a first
approach to address this challenging task, we propose SceneAdapt, a method for
scene adaptation of semantic segmentation algorithms based on adversarial
learning. Experiments and comparisons with state-of-the-art approaches to
domain adaptation highlight that promising performance can be achieved using
adversarial learning both when the two scenes have different but points of
view, and when they comprise images of completely different scenes. To
encourage research on this topic, we made our code available at our web page:
https://iplab.dmi.unict.it/ParkSmartSceneAdaptation/.
Related papers
- Multi-Modal Domain Adaptation Across Video Scenes for Temporal Video
Grounding [59.599378814835205]
Temporal Video Grounding (TVG) aims to localize the temporal boundary of a specific segment in an untrimmed video based on a given language query.
We introduce a novel AMDA method to adaptively adjust the model's scene-related knowledge by incorporating insights from the target data.
arXiv Detail & Related papers (2023-12-21T07:49:27Z) - Learning Semantic Segmentation with Query Points Supervision on Aerial Images [57.09251327650334]
We present a weakly supervised learning algorithm to train semantic segmentation algorithms.
Our proposed approach performs accurate semantic segmentation and improves efficiency by significantly reducing the cost and time required for manual annotation.
arXiv Detail & Related papers (2023-09-11T14:32:04Z) - Shatter and Gather: Learning Referring Image Segmentation with Text
Supervision [52.46081425504072]
We present a new model that discovers semantic entities in input image and then combines such entities relevant to text query to predict the mask of the referent.
Our method was evaluated on four public benchmarks for referring image segmentation, where it clearly outperformed the existing method for the same task and recent open-vocabulary segmentation models on all the benchmarks.
arXiv Detail & Related papers (2023-08-29T15:39:15Z) - Location-Aware Self-Supervised Transformers [74.76585889813207]
We propose to pretrain networks for semantic segmentation by predicting the relative location of image parts.
We control the difficulty of the task by masking a subset of the reference patch features visible to those of the query.
Our experiments show that this location-aware pretraining leads to representations that transfer competitively to several challenging semantic segmentation benchmarks.
arXiv Detail & Related papers (2022-12-05T16:24:29Z) - CYBORGS: Contrastively Bootstrapping Object Representations by Grounding
in Segmentation [22.89327564484357]
We propose a framework which accomplishes this goal via joint learning of representations and segmentation.
By iterating between these two components, we ground the contrastive updates in segmentation information, and simultaneously improve segmentation throughout pretraining.
arXiv Detail & Related papers (2022-03-17T14:20:05Z) - Learning To Segment Dominant Object Motion From Watching Videos [72.57852930273256]
We envision a simple framework for dominant moving object segmentation that neither requires annotated data to train nor relies on saliency priors or pre-trained optical flow maps.
Inspired by a layered image representation, we introduce a technique to group pixel regions according to their affine parametric motion.
This enables our network to learn segmentation of the dominant foreground object using only RGB image pairs as input for both training and inference.
arXiv Detail & Related papers (2021-11-28T14:51:00Z) - Unsupervised Domain Adaptation for Video Semantic Segmentation [91.30558794056054]
Unsupervised Domain Adaptation for semantic segmentation has gained immense popularity since it can transfer knowledge from simulation to real.
In this work, we present a new video extension of this task, namely Unsupervised Domain Adaptation for Video Semantic approaches.
We show that our proposals significantly outperform previous image-based UDA methods both on image-level (mIoU) and video-level (VPQ) evaluation metrics.
arXiv Detail & Related papers (2021-07-23T07:18:20Z) - Realizing Pixel-Level Semantic Learning in Complex Driving Scenes based
on Only One Annotated Pixel per Class [17.481116352112682]
We propose a new semantic segmentation task under complex driving scenes based on weakly supervised condition.
A three step process is built for pseudo labels generation, which progressively implement optimal feature representation for each category.
Experiments on Cityscapes dataset demonstrate that the proposed method provides a feasible way to solve weakly supervised semantic segmentation task.
arXiv Detail & Related papers (2020-03-10T12:57:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.