Unsupervised Camouflaged Object Segmentation as Domain Adaptation
- URL: http://arxiv.org/abs/2308.04528v1
- Date: Tue, 8 Aug 2023 18:46:16 GMT
- Title: Unsupervised Camouflaged Object Segmentation as Domain Adaptation
- Authors: Yi Zhang, Chengyi Wu
- Abstract summary: In this work, we investigate a new task, namely unsupervised camouflaged object segmentation (UCOS), where the target objects own a common attribute, i.e., camouflage.
We find that the state-of-the-art unsupervised models struggle in adapting UCOS, due to the domain gap between the properties of generic and camouflaged objects.
We formulate the UCOS as a source-free unsupervised domain adaptation task (UCOS-DA), where both source labels and target labels are absent during the whole model training process.
- Score: 5.304455190825436
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Deep learning for unsupervised image segmentation remains challenging due to
the absence of human labels. The common idea is to train a segmentation head,
with the supervision of pixel-wise pseudo-labels generated based on the
representation of self-supervised backbones. By doing so, the model performance
depends much on the distance between the distributions of target datasets and
the pre-training dataset (e.g., ImageNet). In this work, we investigate a new
task, namely unsupervised camouflaged object segmentation (UCOS), where the
target objects own a common rarely-seen attribute, i.e., camouflage.
Unsurprisingly, we find that the state-of-the-art unsupervised models struggle
in adapting UCOS, due to the domain gap between the properties of generic and
camouflaged objects. To this end, we formulate the UCOS as a source-free
unsupervised domain adaptation task (UCOS-DA), where both source labels and
target labels are absent during the whole model training process. Specifically,
we define a source model consisting of self-supervised vision transformers
pre-trained on ImageNet. On the other hand, the target domain includes a simple
linear layer (i.e., our target model) and unlabeled camouflaged objects. We
then design a pipeline for foreground-background-contrastive self-adversarial
domain adaptation, to achieve robust UCOS. As a result, our baseline model
achieves superior segmentation performance when compared with competing
unsupervised models on the UCOS benchmark, with the training set which's scale
is only one tenth of the supervised COS counterpart.
Related papers
- SOHES: Self-supervised Open-world Hierarchical Entity Segmentation [82.45303116125021]
This work presents Self-supervised Open-world Hierarchical Entities (SOHES), a novel approach that eliminates the need for human annotations.
We produce abundant high-quality pseudo-labels through visual feature clustering, and rectify the noises in pseudo-labels via a teacher- mutual-learning procedure.
Using raw images as the sole training data, our method achieves unprecedented performance in self-supervised open-world segmentation.
arXiv Detail & Related papers (2024-04-18T17:59:46Z) - Weakly-supervised Contrastive Learning for Unsupervised Object Discovery [52.696041556640516]
Unsupervised object discovery is promising due to its ability to discover objects in a generic manner.
We design a semantic-guided self-supervised learning model to extract high-level semantic features from images.
We introduce Principal Component Analysis (PCA) to localize object regions.
arXiv Detail & Related papers (2023-07-07T04:03:48Z) - Regularizing Self-training for Unsupervised Domain Adaptation via
Structural Constraints [14.593782939242121]
We propose to incorporate structural cues from auxiliary modalities, such as depth, to regularise conventional self-training objectives.
Specifically, we introduce a contrastive pixel-level objectness constraint that pulls the pixel representations within a region of an object instance closer.
We show that our regularizer significantly improves top performing self-training methods in various UDA benchmarks for semantic segmentation.
arXiv Detail & Related papers (2023-04-29T00:12:26Z) - Location-Aware Self-Supervised Transformers [74.76585889813207]
We propose to pretrain networks for semantic segmentation by predicting the relative location of image parts.
We control the difficulty of the task by masking a subset of the reference patch features visible to those of the query.
Our experiments show that this location-aware pretraining leads to representations that transfer competitively to several challenging semantic segmentation benchmarks.
arXiv Detail & Related papers (2022-12-05T16:24:29Z) - Unsupervised domain adaptation for clinician pose estimation and
instance segmentation in the OR [4.024513066910992]
We study how joint person pose estimation and segmentation instance can be performed on low resolutions images from 1x to 12x.
We propose a novel unsupervised domain adaptation method, called emphAdaptOR, to adapt a model from an emphin-the-wild labeled source domain to a statistically different unlabeled target domain.
We show the generality of our method as a semi-supervised learning (SSL) method on the large-scale emphCOCO dataset.
arXiv Detail & Related papers (2021-08-26T14:07:43Z) - PixMatch: Unsupervised Domain Adaptation via Pixelwise Consistency
Training [4.336877104987131]
Unsupervised domain adaptation is a promising technique for semantic segmentation.
We present a novel framework for unsupervised domain adaptation based on the notion of target-domain consistency training.
Our approach is simpler, easier to implement, and more memory-efficient during training.
arXiv Detail & Related papers (2021-05-17T19:36:28Z) - Semi-Supervised Domain Adaptation with Prototypical Alignment and
Consistency Learning [86.6929930921905]
This paper studies how much it can help address domain shifts if we further have a few target samples labeled.
To explore the full potential of landmarks, we incorporate a prototypical alignment (PA) module which calculates a target prototype for each class from the landmarks.
Specifically, we severely perturb the labeled images, making PA non-trivial to achieve and thus promoting model generalizability.
arXiv Detail & Related papers (2021-04-19T08:46:08Z) - Group-Wise Semantic Mining for Weakly Supervised Semantic Segmentation [49.90178055521207]
This work addresses weakly supervised semantic segmentation (WSSS), with the goal of bridging the gap between image-level annotations and pixel-level segmentation.
We formulate WSSS as a novel group-wise learning task that explicitly models semantic dependencies in a group of images to estimate more reliable pseudo ground-truths.
In particular, we devise a graph neural network (GNN) for group-wise semantic mining, wherein input images are represented as graph nodes.
arXiv Detail & Related papers (2020-12-09T12:40:13Z) - Learning from Scale-Invariant Examples for Domain Adaptation in Semantic
Segmentation [6.320141734801679]
We propose a novel approach of exploiting scale-invariance property of semantic segmentation model for self-supervised domain adaptation.
Our algorithm is based on a reasonable assumption that, in general, regardless of the size of the object and stuff (given context) the semantic labeling should be unchanged.
We show that this constraint is violated over the images of the target domain, and hence could be used to transfer labels in-between differently scaled patches.
arXiv Detail & Related papers (2020-07-28T19:40:45Z) - Alleviating Semantic-level Shift: A Semi-supervised Domain Adaptation
Method for Semantic Segmentation [97.8552697905657]
A key challenge of this task is how to alleviate the data distribution discrepancy between the source and target domains.
We propose Alleviating Semantic-level Shift (ASS), which can successfully promote the distribution consistency from both global and local views.
We apply our ASS to two domain adaptation tasks, from GTA5 to Cityscapes and from Synthia to Cityscapes.
arXiv Detail & Related papers (2020-04-02T03:25:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.