DaMSTF: Domain Adversarial Learning Enhanced Meta Self-Training for
Domain Adaptation
- URL: http://arxiv.org/abs/2308.02753v1
- Date: Sat, 5 Aug 2023 00:14:49 GMT
- Title: DaMSTF: Domain Adversarial Learning Enhanced Meta Self-Training for
Domain Adaptation
- Authors: Menglong Lu, Zhen Huang, Yunxiang Zhao, Zhiliang Tian, Yang Liu and
Dongsheng Li
- Abstract summary: We propose a new self-training framework for domain adaptation, namely Domain adversarial learning enhanced Self-Training Framework (DaMSTF)
DaMSTF involves meta-learning to estimate the importance of each pseudo instance, so as to simultaneously reduce the label noise and preserve hard examples.
DaMSTF improves the performance of BERT with an average of nearly 4%.
- Score: 20.697905456202754
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Self-training emerges as an important research line on domain adaptation. By
taking the model's prediction as the pseudo labels of the unlabeled data,
self-training bootstraps the model with pseudo instances in the target domain.
However, the prediction errors of pseudo labels (label noise) challenge the
performance of self-training. To address this problem, previous approaches only
use reliable pseudo instances, i.e., pseudo instances with high prediction
confidence, to retrain the model. Although these strategies effectively reduce
the label noise, they are prone to miss the hard examples. In this paper, we
propose a new self-training framework for domain adaptation, namely Domain
adversarial learning enhanced Self-Training Framework (DaMSTF). Firstly, DaMSTF
involves meta-learning to estimate the importance of each pseudo instance, so
as to simultaneously reduce the label noise and preserve hard examples.
Secondly, we design a meta constructor for constructing the meta-validation
set, which guarantees the effectiveness of the meta-learning module by
improving the quality of the meta-validation set. Thirdly, we find that the
meta-learning module suffers from the training guidance vanishment and tends to
converge to an inferior optimal. To this end, we employ domain adversarial
learning as a heuristic neural network initialization method, which can help
the meta-learning module converge to a better optimal. Theoretically and
experimentally, we demonstrate the effectiveness of the proposed DaMSTF. On the
cross-domain sentiment classification task, DaMSTF improves the performance of
BERT with an average of nearly 4%.
Related papers
- Unsupervised Domain Adaptation for Semantic Segmentation with Pseudo
Label Self-Refinement [9.69089112870202]
We propose an auxiliary pseudo-label refinement network (PRN) for online refining of the pseudo labels and also localizing the pixels whose predicted labels are likely to be noisy.
We evaluate our approach on benchmark datasets with three different domain shifts, and our approach consistently performs significantly better than the previous state-of-the-art methods.
arXiv Detail & Related papers (2023-10-25T20:31:07Z) - Informative Data Mining for One-Shot Cross-Domain Semantic Segmentation [84.82153655786183]
We propose a novel framework called Informative Data Mining (IDM) to enable efficient one-shot domain adaptation for semantic segmentation.
IDM provides an uncertainty-based selection criterion to identify the most informative samples, which facilitates quick adaptation and reduces redundant training.
Our approach outperforms existing methods and achieves a new state-of-the-art one-shot performance of 56.7%/55.4% on the GTA5/SYNTHIA to Cityscapes adaptation tasks.
arXiv Detail & Related papers (2023-09-25T15:56:01Z) - Robust Target Training for Multi-Source Domain Adaptation [110.77704026569499]
We propose a novel Bi-level Optimization based Robust Target Training (BORT$2$) method for MSDA.
Our proposed method achieves the state of the art performance on three MSDA benchmarks, including the large-scale DomainNet dataset.
arXiv Detail & Related papers (2022-10-04T15:20:01Z) - Adapting the Mean Teacher for keypoint-based lung registration under
geometric domain shifts [75.51482952586773]
deep neural networks generally require plenty of labeled training data and are vulnerable to domain shifts between training and test data.
We present a novel approach to geometric domain adaptation for image registration, adapting a model from a labeled source to an unlabeled target domain.
Our method consistently improves on the baseline model by 50%/47% while even matching the accuracy of models trained on target data.
arXiv Detail & Related papers (2022-07-01T12:16:42Z) - Boosting Cross-Domain Speech Recognition with Self-Supervision [35.01508881708751]
Cross-domain performance of automatic speech recognition (ASR) could be severely hampered due to mismatch between training and testing distributions.
Previous work has shown that self-supervised learning (SSL) or pseudo-labeling (PL) is effective in UDA by exploiting the self-supervisions of unlabeled data.
This work presents a systematic UDA framework to fully utilize the unlabeled data with self-supervision in the pre-training and fine-tuning paradigm.
arXiv Detail & Related papers (2022-06-20T14:02:53Z) - Source-Free Open Compound Domain Adaptation in Semantic Segmentation [99.82890571842603]
In SF-OCDA, only the source pre-trained model and the target data are available to learn the target model.
We propose the Cross-Patch Style Swap (CPSS) to diversify samples with various patch styles in the feature-level.
Our method produces state-of-the-art results on the C-Driving dataset.
arXiv Detail & Related papers (2021-06-07T08:38:41Z) - Contrastive Learning and Self-Training for Unsupervised Domain
Adaptation in Semantic Segmentation [71.77083272602525]
UDA attempts to provide efficient knowledge transfer from a labeled source domain to an unlabeled target domain.
We propose a contrastive learning approach that adapts category-wise centroids across domains.
We extend our method with self-training, where we use a memory-efficient temporal ensemble to generate consistent and reliable pseudo-labels.
arXiv Detail & Related papers (2021-05-05T11:55:53Z) - Adaptive Boosting for Domain Adaptation: Towards Robust Predictions in
Scene Segmentation [41.05407168312345]
Domain adaptation is to transfer the shared knowledge learned from the source domain to a new environment, i.e., target domain.
One common practice is to train the model on both labeled source-domain data and unlabeled target-domain data.
We propose one efficient bootstrapping method, called Adaboost Student, explicitly learning complementary models during training.
arXiv Detail & Related papers (2021-03-29T15:12:58Z) - Two-phase Pseudo Label Densification for Self-training based Domain
Adaptation [93.03265290594278]
We propose a novel Two-phase Pseudo Label Densification framework, referred to as TPLD.
In the first phase, we use sliding window voting to propagate the confident predictions, utilizing intrinsic spatial-correlations in the images.
In the second phase, we perform a confidence-based easy-hard classification.
To ease the training process and avoid noisy predictions, we introduce the bootstrapping mechanism to the original self-training loss.
arXiv Detail & Related papers (2020-12-09T02:35:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.