AdaTriplet-RA: Domain Matching via Adaptive Triplet and Reinforced
Attention for Unsupervised Domain Adaptation
- URL: http://arxiv.org/abs/2211.08894v1
- Date: Wed, 16 Nov 2022 13:04:24 GMT
- Title: AdaTriplet-RA: Domain Matching via Adaptive Triplet and Reinforced
Attention for Unsupervised Domain Adaptation
- Authors: Xinyao Shu, Shiyang Yan, Zhenyu Lu, Xinshao Wang, Yuan Xie
- Abstract summary: Unsupervised domain adaption (UDA) is a transfer learning task where the data and annotations of the source domain are available but only have access to the unlabeled target data during training.
We propose to improve the unsupervised domain adaptation task with an inter-domain sample matching scheme.
We apply the widely-used and robust Triplet loss to match the inter-domain samples.
To reduce the catastrophic effect of the inaccurate pseudo-labels generated during training, we propose a novel uncertainty measurement method to select reliable pseudo-labels automatically and progressively refine them.
- Score: 15.905869933337101
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Unsupervised domain adaption (UDA) is a transfer learning task where the data
and annotations of the source domain are available but only have access to the
unlabeled target data during training. Most previous methods try to minimise
the domain gap by performing distribution alignment between the source and
target domains, which has a notable limitation, i.e., operating at the domain
level, but neglecting the sample-level differences. To mitigate this weakness,
we propose to improve the unsupervised domain adaptation task with an
inter-domain sample matching scheme. We apply the widely-used and robust
Triplet loss to match the inter-domain samples. To reduce the catastrophic
effect of the inaccurate pseudo-labels generated during training, we propose a
novel uncertainty measurement method to select reliable pseudo-labels
automatically and progressively refine them. We apply the advanced discrete
relaxation Gumbel Softmax technique to realise an adaptive Topk scheme to
fulfil the functionality. In addition, to enable the global ranking
optimisation within one batch for the domain matching, the whole model is
optimised via a novel reinforced attention mechanism with supervision from the
policy gradient algorithm, using the Average Precision (AP) as the reward. Our
model (termed \textbf{\textit{AdaTriplet-RA}}) achieves State-of-the-art
results on several public benchmark datasets, and its effectiveness is
validated via comprehensive ablation studies. Our method improves the accuracy
of the baseline by 9.7\% (ResNet-101) and 6.2\% (ResNet-50) on the VisDa
dataset and 4.22\% (ResNet-50) on the Domainnet dataset. {The source code is
publicly available at \textit{https://github.com/shuxy0120/AdaTriplet-RA}}.
Related papers
- Divide and Contrast: Source-free Domain Adaptation via Adaptive
Contrastive Learning [122.62311703151215]
Divide and Contrast (DaC) aims to connect the good ends of both worlds while bypassing their limitations.
DaC divides the target data into source-like and target-specific samples, where either group of samples is treated with tailored goals.
We further align the source-like domain with the target-specific samples using a memory bank-based Maximum Mean Discrepancy (MMD) loss to reduce the distribution mismatch.
arXiv Detail & Related papers (2022-11-12T09:21:49Z) - Robust Target Training for Multi-Source Domain Adaptation [110.77704026569499]
We propose a novel Bi-level Optimization based Robust Target Training (BORT$2$) method for MSDA.
Our proposed method achieves the state of the art performance on three MSDA benchmarks, including the large-scale DomainNet dataset.
arXiv Detail & Related papers (2022-10-04T15:20:01Z) - Cycle Label-Consistent Networks for Unsupervised Domain Adaptation [57.29464116557734]
Domain adaptation aims to leverage a labeled source domain to learn a classifier for the unlabeled target domain with a different distribution.
We propose a simple yet efficient domain adaptation method, i.e. Cycle Label-Consistent Network (CLCN), by exploiting the cycle consistency of classification label.
We demonstrate the effectiveness of our approach on MNIST-USPS-SVHN, Office-31, Office-Home and Image CLEF-DA benchmarks.
arXiv Detail & Related papers (2022-05-27T13:09:08Z) - Instance Level Affinity-Based Transfer for Unsupervised Domain
Adaptation [74.71931918541748]
We propose an instance affinity based criterion for source to target transfer during adaptation, called ILA-DA.
We first propose a reliable and efficient method to extract similar and dissimilar samples across source and target, and utilize a multi-sample contrastive loss to drive the domain alignment process.
We verify the effectiveness of ILA-DA by observing consistent improvements in accuracy over popular domain adaptation approaches on a variety of benchmark datasets.
arXiv Detail & Related papers (2021-04-03T01:33:14Z) - Surprisingly Simple Semi-Supervised Domain Adaptation with Pretraining
and Consistency [93.89773386634717]
Visual domain adaptation involves learning to classify images from a target visual domain using labels available in a different source domain.
We show that in the presence of a few target labels, simple techniques like self-supervision (via rotation prediction) and consistency regularization can be effective without any adversarial alignment to learn a good target classifier.
Our Pretraining and Consistency (PAC) approach, can achieve state of the art accuracy on this semi-supervised domain adaptation task, surpassing multiple adversarial domain alignment methods, across multiple datasets.
arXiv Detail & Related papers (2021-01-29T18:40:17Z) - A Free Lunch for Unsupervised Domain Adaptive Object Detection without
Source Data [69.091485888121]
Unsupervised domain adaptation assumes that source and target domain data are freely available and usually trained together to reduce the domain gap.
We propose a source data-free domain adaptive object detection (SFOD) framework via modeling it into a problem of learning with noisy labels.
arXiv Detail & Related papers (2020-12-10T01:42:35Z) - Inductive Unsupervised Domain Adaptation for Few-Shot Classification via
Clustering [16.39667909141402]
Few-shot classification tends to struggle when it needs to adapt to diverse domains.
We introduce a framework, DaFeC, to improve Domain adaptation performance for Few-shot classification via Clustering.
Our approach outperforms previous work with absolute gains (in classification accuracy) of 4.95%, 9.55%, 3.99% and 11.62%, respectively.
arXiv Detail & Related papers (2020-06-23T08:17:48Z) - A simple baseline for domain adaptation using rotation prediction [17.539027866457673]
The goal is to adapt a model trained in one domain to another domain with scarce annotated data.
We propose a simple yet effective method based on self-supervised learning.
Our simple method achieves state-of-the-art results on semi-supervised domain adaptation on DomainNet dataset.
arXiv Detail & Related papers (2019-12-26T17:32:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.