Domain Impression: A Source Data Free Domain Adaptation Method
- URL: http://arxiv.org/abs/2102.09003v1
- Date: Wed, 17 Feb 2021 19:50:49 GMT
- Title: Domain Impression: A Source Data Free Domain Adaptation Method
- Authors: Vinod K Kurmi and Venkatesh K Subramanian and Vinay P Namboodiri
- Abstract summary: Unsupervised domain adaptation methods solve the adaptation problem for an unlabeled target set, assuming that the source dataset is available with all labels.
This paper proposes a domain adaptation technique that does not need any source data.
Instead of the source data, we are only provided with a classifier that is trained on the source data.
- Score: 27.19677042654432
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Unsupervised Domain adaptation methods solve the adaptation problem for an
unlabeled target set, assuming that the source dataset is available with all
labels. However, the availability of actual source samples is not always
possible in practical cases. It could be due to memory constraints, privacy
concerns, and challenges in sharing data. This practical scenario creates a
bottleneck in the domain adaptation problem. This paper addresses this
challenging scenario by proposing a domain adaptation technique that does not
need any source data. Instead of the source data, we are only provided with a
classifier that is trained on the source data. Our proposed approach is based
on a generative framework, where the trained classifier is used for generating
samples from the source classes. We learn the joint distribution of data by
using the energy-based modeling of the trained classifier. At the same time, a
new classifier is also adapted for the target domain. We perform various
ablation analysis under different experimental setups and demonstrate that the
proposed approach achieves better results than the baseline models in this
extremely novel scenario.
Related papers
- Divide and Contrast: Source-free Domain Adaptation via Adaptive
Contrastive Learning [122.62311703151215]
Divide and Contrast (DaC) aims to connect the good ends of both worlds while bypassing their limitations.
DaC divides the target data into source-like and target-specific samples, where either group of samples is treated with tailored goals.
We further align the source-like domain with the target-specific samples using a memory bank-based Maximum Mean Discrepancy (MMD) loss to reduce the distribution mismatch.
arXiv Detail & Related papers (2022-11-12T09:21:49Z) - Source-Free Domain Adaptation via Distribution Estimation [106.48277721860036]
Domain Adaptation aims to transfer the knowledge learned from a labeled source domain to an unlabeled target domain whose data distributions are different.
Recently, Source-Free Domain Adaptation (SFDA) has drawn much attention, which tries to tackle domain adaptation problem without using source data.
In this work, we propose a novel framework called SFDA-DE to address SFDA task via source Distribution Estimation.
arXiv Detail & Related papers (2022-04-24T12:22:19Z) - Unsupervised Adaptation of Semantic Segmentation Models without Source
Data [14.66682099621276]
We consider the novel problem of unsupervised domain adaptation of source models, without access to the source data for semantic segmentation.
We propose a self-training approach to extract the knowledge from the source model.
Our framework is able to achieve significant performance gains compared to directly applying the source model on the target data.
arXiv Detail & Related papers (2021-12-04T15:13:41Z) - Distill and Fine-tune: Effective Adaptation from a Black-box Source
Model [138.12678159620248]
Unsupervised domain adaptation (UDA) aims to transfer knowledge in previous related labeled datasets (source) to a new unlabeled dataset (target)
We propose a novel two-step adaptation framework called Distill and Fine-tune (Dis-tune)
arXiv Detail & Related papers (2021-04-04T05:29:05Z) - Source-free Domain Adaptation via Distributional Alignment by Matching
Batch Normalization Statistics [85.75352990739154]
We propose a novel domain adaptation method for the source-free setting.
We use batch normalization statistics stored in the pretrained model to approximate the distribution of unobserved source data.
Our method achieves competitive performance with state-of-the-art domain adaptation methods.
arXiv Detail & Related papers (2021-01-19T14:22:33Z) - Open-Set Hypothesis Transfer with Semantic Consistency [99.83813484934177]
We introduce a method that focuses on the semantic consistency under transformation of target data.
Our model first discovers confident predictions and performs classification with pseudo-labels.
As a result, unlabeled data can be classified into discriminative classes coincided with either source classes or unknown classes.
arXiv Detail & Related papers (2020-10-01T10:44:31Z) - Unsupervised Domain Adaptation in the Absence of Source Data [0.7366405857677227]
We propose an unsupervised method for adapting a source classifier to a target domain that varies from the source domain along natural axes.
We validate our method in scenarios where the distribution shift involves brightness, contrast, and rotation and show that it outperforms fine-tuning baselines in scenarios with limited labeled data.
arXiv Detail & Related papers (2020-07-20T16:22:14Z) - Do We Really Need to Access the Source Data? Source Hypothesis Transfer
for Unsupervised Domain Adaptation [102.67010690592011]
Unsupervised adaptationUDA (UDA) aims to leverage the knowledge learned from a labeled source dataset to solve similar tasks in a new unlabeled domain.
Prior UDA methods typically require to access the source data when learning to adapt the model.
This work tackles a practical setting where only a trained source model is available and how we can effectively utilize such a model without source data to solve UDA problems.
arXiv Detail & Related papers (2020-02-20T03:13:58Z) - Domain Adaptive Bootstrap Aggregating [5.444459446244819]
bootstrap aggregating, or bagging, is a popular method for improving stability of predictive algorithms.
This article proposes a domain adaptive bagging method coupled with a new iterative nearest neighbor sampler.
arXiv Detail & Related papers (2020-01-12T20:02:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.