WEDGE: Web-Image Assisted Domain Generalization for Semantic
Segmentation
- URL: http://arxiv.org/abs/2109.14196v4
- Date: Tue, 2 May 2023 05:59:19 GMT
- Title: WEDGE: Web-Image Assisted Domain Generalization for Semantic
Segmentation
- Authors: Namyup Kim, Taeyoung Son, Jaehyun Pahk, Cuiling Lan, Wenjun Zeng, Suha
Kwak
- Abstract summary: We propose a WEb-image assisted Domain GEneralization scheme, which is the first to exploit the diversity of web-crawled images for generalizable semantic segmentation.
We also present a method which injects styles of the web-crawled images into training images on-the-fly during training, which enables the network to experience images of diverse styles with reliable labels for effective training.
- Score: 72.88657378658549
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Domain generalization for semantic segmentation is highly demanded in real
applications, where a trained model is expected to work well in previously
unseen domains. One challenge lies in the lack of data which could cover the
diverse distributions of the possible unseen domains for training. In this
paper, we propose a WEb-image assisted Domain GEneralization (WEDGE) scheme,
which is the first to exploit the diversity of web-crawled images for
generalizable semantic segmentation. To explore and exploit the real-world data
distributions, we collect web-crawled images which present large diversity in
terms of weather conditions, sites, lighting, camera styles, etc. We also
present a method which injects styles of the web-crawled images into training
images on-the-fly during training, which enables the network to experience
images of diverse styles with reliable labels for effective training. Moreover,
we use the web-crawled images with their predicted pseudo labels for training
to further enhance the capability of the network. Extensive experiments
demonstrate that our method clearly outperforms existing domain generalization
techniques.
Related papers
- Learning from the Web: Language Drives Weakly-Supervised Incremental Learning for Semantic Segmentation [33.955384040748946]
We argue that widely available web images can also be considered for the learning of new classes.
To our knowledge, this is the first work to rely solely on web images for both the learning of new concepts and the preservation of the already learned ones.
arXiv Detail & Related papers (2024-07-18T10:14:49Z) - WIDIn: Wording Image for Domain-Invariant Representation in Single-Source Domain Generalization [63.98650220772378]
We present WIDIn, Wording Images for Domain-Invariant representation, to disentangle discriminative visual representation.
We first estimate the language embedding with fine-grained alignment, which can be used to adaptively identify and then remove domain-specific counterpart.
We show that WIDIn can be applied to both pretrained vision-language models like CLIP, and separately trained uni-modal models like MoCo and BERT.
arXiv Detail & Related papers (2024-05-28T17:46:27Z) - ODES: Domain Adaptation with Expert Guidance for Online Medical Image Segmentation [16.90507882617707]
Domain adaptive segmentation typically relies on self-training using pseudo labels predicted by a pre-trained network on an unlabeled target dataset.
The acquisition of pixel-level annotations across all images in a batch often leads to redundant information while increasing temporal overhead in online learning.
We propose a novel image-pruning strategy that selects the most useful subset of images from the current batch for active learning.
arXiv Detail & Related papers (2023-12-08T23:43:17Z) - Domain-Controlled Prompt Learning [49.45309818782329]
Existing prompt learning methods often lack domain-awareness or domain-transfer mechanisms.
We propose a textbfDomain-Controlled Prompt Learning for the specific domains.
Our method achieves state-of-the-art performance in specific domain image recognition datasets.
arXiv Detail & Related papers (2023-09-30T02:59:49Z) - Using Language to Extend to Unseen Domains [81.37175826824625]
It is expensive to collect training data for every possible domain that a vision model may encounter when deployed.
We consider how simply verbalizing the training domain as well as domains we want to extend to but do not have data for can improve robustness.
Using a multimodal model with a joint image and language embedding space, our method LADS learns a transformation of the image embeddings from the training domain to each unseen test domain.
arXiv Detail & Related papers (2022-10-18T01:14:02Z) - Federated Domain Generalization for Image Recognition via Cross-Client
Style Transfer [60.70102634957392]
Domain generalization (DG) has been a hot topic in image recognition, with a goal to train a general model that can perform well on unseen domains.
In this paper, we propose a novel domain generalization method for image recognition through cross-client style transfer (CCST) without exchanging data samples.
Our method outperforms recent SOTA DG methods on two DG benchmarks (PACS, OfficeHome) and a large-scale medical image dataset (Camelyon17) in the FL setting.
arXiv Detail & Related papers (2022-10-03T13:15:55Z) - Domain-invariant Prototypes for Semantic Segmentation [30.932130453313537]
We present an easy-to-train framework that learns domain-invariant prototypes for domain adaptive semantic segmentation.
Our method involves only one-stage training and does not need to be trained on large-scale un-annotated target images.
arXiv Detail & Related papers (2022-08-12T02:21:05Z) - DoFE: Domain-oriented Feature Embedding for Generalizable Fundus Image
Segmentation on Unseen Datasets [96.92018649136217]
We present a novel Domain-oriented Feature Embedding (DoFE) framework to improve the generalization ability of CNNs on unseen target domains.
Our DoFE framework dynamically enriches the image features with additional domain prior knowledge learned from multi-source domains.
Our framework generates satisfying segmentation results on unseen datasets and surpasses other domain generalization and network regularization methods.
arXiv Detail & Related papers (2020-10-13T07:28:39Z) - Generalizable Model-agnostic Semantic Segmentation via Target-specific
Normalization [24.14272032117714]
We propose a novel domain generalization framework for the generalizable semantic segmentation task.
We exploit the model-agnostic learning to simulate the domain shift problem.
Considering the data-distribution discrepancy between seen source and unseen target domains, we develop the target-specific normalization scheme.
arXiv Detail & Related papers (2020-03-27T09:25:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.