DomainInv: Domain Invariant Fine Tuning and Adversarial Label Correction
For QA Domain Adaptation
- URL: http://arxiv.org/abs/2305.05589v1
- Date: Thu, 4 May 2023 18:13:17 GMT
- Title: DomainInv: Domain Invariant Fine Tuning and Adversarial Label Correction
For QA Domain Adaptation
- Authors: Anant Khandelwal
- Abstract summary: Existing Question Answering (QA) systems limited by the capability of answering questions from unseen domain or any out-of-domain distributions.
Most importantly all the existing QA domain adaptation methods are either based on generating synthetic data or pseudo labeling the target domain data.
In this paper, we propose the unsupervised domain adaptation for unlabeled target domain by transferring the target representation near to source domain while still using the supervision from source domain.
- Score: 27.661609140918916
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Existing Question Answering (QA) systems limited by the capability of
answering questions from unseen domain or any out-of-domain distributions
making them less reliable for deployment to real scenarios. Most importantly
all the existing QA domain adaptation methods are either based on generating
synthetic data or pseudo labeling the target domain data. The domain adaptation
methods based on synthetic data and pseudo labeling suffers either from the
requirement of computational resources or an extra overhead of carefully
selecting the confidence threshold to separate the noisy examples from being in
the training dataset. In this paper, we propose the unsupervised domain
adaptation for unlabeled target domain by transferring the target
representation near to source domain while still using the supervision from
source domain. Towards that we proposed the idea of domain invariant fine
tuning along with adversarial label correction to identify the target instances
which lie far apart from the source domain, so that the feature encoder can be
learnt to minimize the distance between such target instances and source
instances class wisely, removing the possibility of learning the features of
target domain which are still near to source support but are ambiguous.
Evaluation of our QA domain adaptation method namely, DomainInv on multiple
target QA dataset reveal the performance improvement over the strongest
baseline.
Related papers
- Contrastive Adversarial Training for Unsupervised Domain Adaptation [2.432037584128226]
Domain adversarial training has been successfully adopted for various domain adaptation tasks.
Large models make adversarial training being easily biased towards source domain and hardly adapted to target domain.
We propose contrastive adversarial training (CAT) approach that leverages the labeled source domain samples to reinforce and regulate the feature generation for target domain.
arXiv Detail & Related papers (2024-07-17T17:59:21Z) - CA-UDA: Class-Aware Unsupervised Domain Adaptation with Optimal
Assignment and Pseudo-Label Refinement [84.10513481953583]
unsupervised domain adaptation (UDA) focuses on the selection of good pseudo-labels as surrogates for the missing labels in the target data.
source domain bias that deteriorates the pseudo-labels can still exist since the shared network of the source and target domains are typically used for the pseudo-label selections.
We propose CA-UDA to improve the quality of the pseudo-labels and UDA results with optimal assignment, a pseudo-label refinement strategy and class-aware domain alignment.
arXiv Detail & Related papers (2022-05-26T18:45:04Z) - Instance Level Affinity-Based Transfer for Unsupervised Domain
Adaptation [74.71931918541748]
We propose an instance affinity based criterion for source to target transfer during adaptation, called ILA-DA.
We first propose a reliable and efficient method to extract similar and dissimilar samples across source and target, and utilize a multi-sample contrastive loss to drive the domain alignment process.
We verify the effectiveness of ILA-DA by observing consistent improvements in accuracy over popular domain adaptation approaches on a variety of benchmark datasets.
arXiv Detail & Related papers (2021-04-03T01:33:14Z) - Disentanglement-based Cross-Domain Feature Augmentation for Effective
Unsupervised Domain Adaptive Person Re-identification [87.72851934197936]
Unsupervised domain adaptive (UDA) person re-identification (ReID) aims to transfer the knowledge from the labeled source domain to the unlabeled target domain for person matching.
One challenge is how to generate target domain samples with reliable labels for training.
We propose a Disentanglement-based Cross-Domain Feature Augmentation strategy.
arXiv Detail & Related papers (2021-03-25T15:28:41Z) - Surprisingly Simple Semi-Supervised Domain Adaptation with Pretraining
and Consistency [93.89773386634717]
Visual domain adaptation involves learning to classify images from a target visual domain using labels available in a different source domain.
We show that in the presence of a few target labels, simple techniques like self-supervision (via rotation prediction) and consistency regularization can be effective without any adversarial alignment to learn a good target classifier.
Our Pretraining and Consistency (PAC) approach, can achieve state of the art accuracy on this semi-supervised domain adaptation task, surpassing multiple adversarial domain alignment methods, across multiple datasets.
arXiv Detail & Related papers (2021-01-29T18:40:17Z) - Learning Target Domain Specific Classifier for Partial Domain Adaptation [85.71584004185031]
Unsupervised domain adaptation (UDA) aims at reducing the distribution discrepancy when transferring knowledge from a labeled source domain to an unlabeled target domain.
This paper focuses on a more realistic UDA scenario, where the target label space is subsumed to the source label space.
arXiv Detail & Related papers (2020-08-25T02:28:24Z) - Cross-domain Self-supervised Learning for Domain Adaptation with Few
Source Labels [78.95901454696158]
We propose a novel Cross-Domain Self-supervised learning approach for domain adaptation.
Our method significantly boosts performance of target accuracy in the new target domain with few source labels.
arXiv Detail & Related papers (2020-03-18T15:11:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.