Unsupervised Multi-source Domain Adaptation Without Access to Source
Data
- URL: http://arxiv.org/abs/2104.01845v1
- Date: Mon, 5 Apr 2021 10:45:12 GMT
- Title: Unsupervised Multi-source Domain Adaptation Without Access to Source
Data
- Authors: Sk Miraj Ahmed, Dripta S. Raychaudhuri, Sujoy Paul, Samet Oymak, Amit
K. Roy-Chowdhury
- Abstract summary: Unsupervised Domain Adaptation (UDA) aims to learn a predictor model for an unlabeled domain by transferring knowledge from a separate labeled source domain.
We propose a novel and efficient algorithm which automatically combines the source models with suitable weights in such a way that it performs at least as good as the best source model.
- Score: 58.551861130011886
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Unsupervised Domain Adaptation (UDA) aims to learn a predictor model for an
unlabeled domain by transferring knowledge from a separate labeled source
domain. However, most of these conventional UDA approaches make the strong
assumption of having access to the source data during training, which may not
be very practical due to privacy, security and storage concerns. A recent line
of work addressed this problem and proposed an algorithm that transfers
knowledge to the unlabeled target domain from a single source model without
requiring access to the source data. However, for adaptation purposes, if there
are multiple trained source models available to choose from, this method has to
go through adapting each and every model individually, to check for the best
source. Thus, we ask the question: can we find the optimal combination of
source models, with no source data and without target labels, whose performance
is no worse than the single best source? To answer this, we propose a novel and
efficient algorithm which automatically combines the source models with
suitable weights in such a way that it performs at least as good as the best
source model. We provide intuitive theoretical insights to justify our claim.
Furthermore, extensive experiments are conducted on several benchmark datasets
to show the effectiveness of our algorithm, where in most cases, our method not
only reaches best source accuracy but also outperforms it.
Related papers
- Source-Free Domain-Invariant Performance Prediction [68.39031800809553]
We propose a source-free approach centred on uncertainty-based estimation, using a generative model for calibration in the absence of source data.
Our experiments on benchmark object recognition datasets reveal that existing source-based methods fall short with limited source sample availability.
Our approach significantly outperforms the current state-of-the-art source-free and source-based methods, affirming its effectiveness in domain-invariant performance estimation.
arXiv Detail & Related papers (2024-08-05T03:18:58Z) - Source-Free Domain Adaptation via Distribution Estimation [106.48277721860036]
Domain Adaptation aims to transfer the knowledge learned from a labeled source domain to an unlabeled target domain whose data distributions are different.
Recently, Source-Free Domain Adaptation (SFDA) has drawn much attention, which tries to tackle domain adaptation problem without using source data.
In this work, we propose a novel framework called SFDA-DE to address SFDA task via source Distribution Estimation.
arXiv Detail & Related papers (2022-04-24T12:22:19Z) - Unsupervised Adaptation of Semantic Segmentation Models without Source
Data [14.66682099621276]
We consider the novel problem of unsupervised domain adaptation of source models, without access to the source data for semantic segmentation.
We propose a self-training approach to extract the knowledge from the source model.
Our framework is able to achieve significant performance gains compared to directly applying the source model on the target data.
arXiv Detail & Related papers (2021-12-04T15:13:41Z) - Distill and Fine-tune: Effective Adaptation from a Black-box Source
Model [138.12678159620248]
Unsupervised domain adaptation (UDA) aims to transfer knowledge in previous related labeled datasets (source) to a new unlabeled dataset (target)
We propose a novel two-step adaptation framework called Distill and Fine-tune (Dis-tune)
arXiv Detail & Related papers (2021-04-04T05:29:05Z) - Self-Supervised Noisy Label Learning for Source-Free Unsupervised Domain
Adaptation [87.60688582088194]
We propose a novel Self-Supervised Noisy Label Learning method.
Our method can easily achieve state-of-the-art results and surpass other methods by a very large margin.
arXiv Detail & Related papers (2021-02-23T10:51:45Z) - Unsupervised Domain Adaptation in the Absence of Source Data [0.7366405857677227]
We propose an unsupervised method for adapting a source classifier to a target domain that varies from the source domain along natural axes.
We validate our method in scenarios where the distribution shift involves brightness, contrast, and rotation and show that it outperforms fine-tuning baselines in scenarios with limited labeled data.
arXiv Detail & Related papers (2020-07-20T16:22:14Z) - Multi-source Attention for Unsupervised Domain Adaptation [15.900069711477542]
We model source-selection as an attention-learning problem, where we learn attention over sources for a given target instance.
For this purpose, we first independently learn source-specific classification models, and a relatedness map between sources and target domains.
We then learn attention-weights over the sources for aggregating the predictions of the source-specific models.
arXiv Detail & Related papers (2020-04-14T15:51:02Z) - Do We Really Need to Access the Source Data? Source Hypothesis Transfer
for Unsupervised Domain Adaptation [102.67010690592011]
Unsupervised adaptationUDA (UDA) aims to leverage the knowledge learned from a labeled source dataset to solve similar tasks in a new unlabeled domain.
Prior UDA methods typically require to access the source data when learning to adapt the model.
This work tackles a practical setting where only a trained source model is available and how we can effectively utilize such a model without source data to solve UDA problems.
arXiv Detail & Related papers (2020-02-20T03:13:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.