Semantic-aware Message Broadcasting for Efficient Unsupervised Domain
Adaptation
- URL: http://arxiv.org/abs/2212.02739v1
- Date: Tue, 6 Dec 2022 04:09:47 GMT
- Title: Semantic-aware Message Broadcasting for Efficient Unsupervised Domain
Adaptation
- Authors: Xin Li, Cuiling Lan, Guoqiang Wei, Zhibo Chen
- Abstract summary: We propose a novel method, Semantic-aware Message Broadcasting (SAMB), which enables more informative and flexible feature alignment for unsupervised domain adaptation (UDA)
We introduce a group of learned group tokens as nodes to aggregate the global information from all image tokens.
In this way, our message broadcasting encourages the group tokens to learn more informative and diverse information for effective domain alignment.
- Score: 40.939984198850496
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Vision transformer has demonstrated great potential in abundant vision tasks.
However, it also inevitably suffers from poor generalization capability when
the distribution shift occurs in testing (i.e., out-of-distribution data). To
mitigate this issue, we propose a novel method, Semantic-aware Message
Broadcasting (SAMB), which enables more informative and flexible feature
alignment for unsupervised domain adaptation (UDA). Particularly, we study the
attention module in the vision transformer and notice that the alignment space
using one global class token lacks enough flexibility, where it interacts
information with all image tokens in the same manner but ignores the rich
semantics of different regions. In this paper, we aim to improve the richness
of the alignment features by enabling semantic-aware adaptive message
broadcasting. Particularly, we introduce a group of learned group tokens as
nodes to aggregate the global information from all image tokens, but encourage
different group tokens to adaptively focus on the message broadcasting to
different semantic regions. In this way, our message broadcasting encourages
the group tokens to learn more informative and diverse information for
effective domain alignment. Moreover, we systematically study the effects of
adversarial-based feature alignment (ADA) and pseudo-label based self-training
(PST) on UDA. We find that one simple two-stage training strategy with the
cooperation of ADA and PST can further improve the adaptation capability of the
vision transformer. Extensive experiments on DomainNet, OfficeHome, and
VisDA-2017 demonstrate the effectiveness of our methods for UDA.
Related papers
- PiPa++: Towards Unification of Domain Adaptive Semantic Segmentation via Self-supervised Learning [34.786268652516355]
Unsupervised domain adaptive segmentation aims to improve the segmentation accuracy of models on target domains without relying on labeled data from those domains.
It seeks to align the feature representations of the source domain (where labeled data is available) and the target domain (where only unlabeled data is present)
arXiv Detail & Related papers (2024-07-24T08:53:29Z) - Tell, Don't Show!: Language Guidance Eases Transfer Across Domains in Images and Videos [69.29778009769862]
We introduce LaGTran, a framework that guides robust transfer of discriminative knowledge from labeled source to unlabeled target data with domain gaps.
Motivated by our observation that semantically richer text modality has more favorable transfer properties, we devise a transfer mechanism to use a source-trained text-classifier to generate predictions on the target text descriptions.
Our approach driven by language guidance is surprisingly easy and simple, yet significantly outperforms all prior approaches on challenging datasets like GeoNet and DomainNet.
arXiv Detail & Related papers (2024-03-08T18:58:46Z) - Universal Domain Adaptation for Robust Handling of Distributional Shifts
in NLP [25.4952909342458]
Universal Domain Adaptation (UniDA) has emerged as a novel research area in computer vision.
We propose a benchmark for natural language that offers thorough viewpoints of the model's generalizability and robustness.
arXiv Detail & Related papers (2023-10-23T12:15:25Z) - Prompt Ensemble Self-training for Open-Vocabulary Domain Adaptation [45.02052030837188]
We study open-vocabulary domain adaptation (OVDA), a new unsupervised domain adaptation framework.
We design a Prompt Ensemble Self-training (PEST) technique that exploits the synergy between vision and language.
PEST outperforms the state-of-the-art consistently across 10 image recognition tasks.
arXiv Detail & Related papers (2023-06-29T03:39:35Z) - Self-Supervised Graph Neural Network for Multi-Source Domain Adaptation [51.21190751266442]
Domain adaptation (DA) tries to tackle the scenarios when the test data does not fully follow the same distribution of the training data.
By learning from large-scale unlabeled samples, self-supervised learning has now become a new trend in deep learning.
We propose a novel textbfSelf-textbfSupervised textbfGraph Neural Network (SSG) to enable more effective inter-task information exchange and knowledge sharing.
arXiv Detail & Related papers (2022-04-08T03:37:56Z) - Audio-Adaptive Activity Recognition Across Video Domains [112.46638682143065]
We leverage activity sounds for domain adaptation as they have less variance across domains and can reliably indicate which activities are not happening.
We propose an audio-adaptive encoder and associated learning methods that discriminatively adjust the visual feature representation.
We also introduce the new task of actor shift, with a corresponding audio-visual dataset, to challenge our method with situations where the activity appearance changes dramatically.
arXiv Detail & Related papers (2022-03-27T08:15:20Z) - Semi-Supervised Domain Adaptation with Prototypical Alignment and
Consistency Learning [86.6929930921905]
This paper studies how much it can help address domain shifts if we further have a few target samples labeled.
To explore the full potential of landmarks, we incorporate a prototypical alignment (PA) module which calculates a target prototype for each class from the landmarks.
Specifically, we severely perturb the labeled images, making PA non-trivial to achieve and thus promoting model generalizability.
arXiv Detail & Related papers (2021-04-19T08:46:08Z) - Alleviating Semantic-level Shift: A Semi-supervised Domain Adaptation
Method for Semantic Segmentation [97.8552697905657]
A key challenge of this task is how to alleviate the data distribution discrepancy between the source and target domains.
We propose Alleviating Semantic-level Shift (ASS), which can successfully promote the distribution consistency from both global and local views.
We apply our ASS to two domain adaptation tasks, from GTA5 to Cityscapes and from Synthia to Cityscapes.
arXiv Detail & Related papers (2020-04-02T03:25:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.