Deep Adversarial Domain Adaptation Based on Multi-layer Joint Kernelized
Distance
- URL: http://arxiv.org/abs/2010.05696v1
- Date: Fri, 9 Oct 2020 02:32:48 GMT
- Title: Deep Adversarial Domain Adaptation Based on Multi-layer Joint Kernelized
Distance
- Authors: Sitong Mao, Jiaxin Chen, Xiao Shen, Fu-lai Chung
- Abstract summary: Domain adaptation refers to the learning scenario that a model learned from the source data is applied on the target data.
The distribution discrepancy between source data and target data can substantially affect the adaptation performance.
A deep adversarial domain adaptation model based on a multi-layer joint kernelized distance metric is proposed.
- Score: 30.452492118887182
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Domain adaptation refers to the learning scenario that a model learned from
the source data is applied on the target data which have the same categories
but different distribution. While it has been widely applied, the distribution
discrepancy between source data and target data can substantially affect the
adaptation performance. The problem has been recently addressed by employing
adversarial learning and distinctive adaptation performance has been reported.
In this paper, a deep adversarial domain adaptation model based on a
multi-layer joint kernelized distance metric is proposed. By utilizing the
abstract features extracted from deep networks, the multi-layer joint
kernelized distance (MJKD) between the $j$th target data predicted as the $m$th
category and all the source data of the $m'$th category is computed. Base on
MJKD, a class-balanced selection strategy is utilized in each category to
select target data that are most likely to be classified correctly and treat
them as labeled data using their pseudo labels. Then an adversarial
architecture is used to draw the newly generated labeled training data and the
remaining target data close to each other. In this way, the target data itself
provide valuable information to enhance the domain adaptation. An analysis of
the proposed method is also given and the experimental results demonstrate that
the proposed method can achieve a better performance than a number of
state-of-the-art methods.
Related papers
- Informative Data Mining for One-Shot Cross-Domain Semantic Segmentation [84.82153655786183]
We propose a novel framework called Informative Data Mining (IDM) to enable efficient one-shot domain adaptation for semantic segmentation.
IDM provides an uncertainty-based selection criterion to identify the most informative samples, which facilitates quick adaptation and reduces redundant training.
Our approach outperforms existing methods and achieves a new state-of-the-art one-shot performance of 56.7%/55.4% on the GTA5/SYNTHIA to Cityscapes adaptation tasks.
arXiv Detail & Related papers (2023-09-25T15:56:01Z) - Multi-Source Soft Pseudo-Label Learning with Domain Similarity-based
Weighting for Semantic Segmentation [2.127049691404299]
This paper describes a method of domain adaptive training for semantic segmentation using multiple source datasets.
We propose a soft pseudo-label generation method by integrating predicted object probabilities from multiple source models.
arXiv Detail & Related papers (2023-03-02T05:20:36Z) - Unsupervised Domain Adaptation via Distilled Discriminative Clustering [45.39542287480395]
We re-cast the domain adaptation problem as discriminative clustering of target data.
We propose to jointly train the network using parallel, supervised learning objectives over labeled source data.
We conduct careful ablation studies and extensive experiments on five popular benchmark datasets.
arXiv Detail & Related papers (2023-02-23T13:03:48Z) - Divide and Contrast: Source-free Domain Adaptation via Adaptive
Contrastive Learning [122.62311703151215]
Divide and Contrast (DaC) aims to connect the good ends of both worlds while bypassing their limitations.
DaC divides the target data into source-like and target-specific samples, where either group of samples is treated with tailored goals.
We further align the source-like domain with the target-specific samples using a memory bank-based Maximum Mean Discrepancy (MMD) loss to reduce the distribution mismatch.
arXiv Detail & Related papers (2022-11-12T09:21:49Z) - Source-Free Domain Adaptation via Distribution Estimation [106.48277721860036]
Domain Adaptation aims to transfer the knowledge learned from a labeled source domain to an unlabeled target domain whose data distributions are different.
Recently, Source-Free Domain Adaptation (SFDA) has drawn much attention, which tries to tackle domain adaptation problem without using source data.
In this work, we propose a novel framework called SFDA-DE to address SFDA task via source Distribution Estimation.
arXiv Detail & Related papers (2022-04-24T12:22:19Z) - Conditional Extreme Value Theory for Open Set Video Domain Adaptation [17.474956295874797]
We propose an open-set video domain adaptation approach to mitigate the domain discrepancy between the source and target data.
To alleviate the negative transfer issue, weights computed by the distance from the sample entropy to the threshold are leveraged in adversarial learning.
The proposed method has been thoroughly evaluated on both small-scale and large-scale cross-domain video datasets.
arXiv Detail & Related papers (2021-09-01T10:51:50Z) - Instance Level Affinity-Based Transfer for Unsupervised Domain
Adaptation [74.71931918541748]
We propose an instance affinity based criterion for source to target transfer during adaptation, called ILA-DA.
We first propose a reliable and efficient method to extract similar and dissimilar samples across source and target, and utilize a multi-sample contrastive loss to drive the domain alignment process.
We verify the effectiveness of ILA-DA by observing consistent improvements in accuracy over popular domain adaptation approaches on a variety of benchmark datasets.
arXiv Detail & Related papers (2021-04-03T01:33:14Z) - Source Data-absent Unsupervised Domain Adaptation through Hypothesis
Transfer and Labeling Transfer [137.36099660616975]
Unsupervised adaptation adaptation (UDA) aims to transfer knowledge from a related but different well-labeled source domain to a new unlabeled target domain.
Most existing UDA methods require access to the source data, and thus are not applicable when the data are confidential and not shareable due to privacy concerns.
This paper aims to tackle a realistic setting with only a classification model available trained over, instead of accessing to the source data.
arXiv Detail & Related papers (2020-12-14T07:28:50Z) - Towards Fair Cross-Domain Adaptation via Generative Learning [50.76694500782927]
Domain Adaptation (DA) targets at adapting a model trained over the well-labeled source domain to the unlabeled target domain lying in different distributions.
We develop a novel Generative Few-shot Cross-domain Adaptation (GFCA) algorithm for fair cross-domain classification.
arXiv Detail & Related papers (2020-03-04T23:25:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.