Prior Omission of Dissimilar Source Domain(s) for Cost-Effective
Few-Shot Learning
- URL: http://arxiv.org/abs/2109.05234v1
- Date: Sat, 11 Sep 2021 09:30:59 GMT
- Title: Prior Omission of Dissimilar Source Domain(s) for Cost-Effective
Few-Shot Learning
- Authors: Zezhong Wang, Hongru Wang, Kwan Wai Chung, Jia Zhu, Gabriel Pui Cheong
Fung, Kam-Fai Wong
- Abstract summary: Few-shot slot tagging is an emerging research topic in the field of Natural Language Understanding (NLU)
With sufficient annotated data from source domains, the key challenge is how to train and adapt the model to another target domain which only has few labels.
- Score: 24.647313693814798
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Few-shot slot tagging is an emerging research topic in the field of Natural
Language Understanding (NLU). With sufficient annotated data from source
domains, the key challenge is how to train and adapt the model to another
target domain which only has few labels. Conventional few-shot approaches use
all the data from the source domains without considering inter-domain relations
and implicitly assume each sample in the domain contributes equally. However,
our experiments show that the data distribution bias among different domains
will significantly affect the adaption performance. Moreover, transferring
knowledge from dissimilar domains will even introduce some extra noises so that
affect the performance of models. To tackle this problem, we propose an
effective similarity-based method to select data from the source domains. In
addition, we propose a Shared-Private Network (SP-Net) for the few-shot slot
tagging task. The words from the same class would have some shared features. We
extract those shared features from the limited annotated data on the target
domain and merge them together as the label embedding to help us predict other
unlabelled data on the target domain. The experiment shows that our method
outperforms the state-of-the-art approaches with fewer source data. The result
also proves that some training data from dissimilar sources are redundant and
even negative for the adaption.
Related papers
- Cross-Domain Few-Shot Classification via Inter-Source Stylization [11.008292768447614]
Cross-Domain Few-Shot Classification (CDFSC) is to accurately classify a target dataset with limited labelled data.
This paper proposes a solution that makes use of multiple source domains without the need for additional labeling costs.
arXiv Detail & Related papers (2022-08-17T01:44:32Z) - Source-Free Domain Adaptation via Distribution Estimation [106.48277721860036]
Domain Adaptation aims to transfer the knowledge learned from a labeled source domain to an unlabeled target domain whose data distributions are different.
Recently, Source-Free Domain Adaptation (SFDA) has drawn much attention, which tries to tackle domain adaptation problem without using source data.
In this work, we propose a novel framework called SFDA-DE to address SFDA task via source Distribution Estimation.
arXiv Detail & Related papers (2022-04-24T12:22:19Z) - Inferring Latent Domains for Unsupervised Deep Domain Adaptation [54.963823285456925]
Unsupervised Domain Adaptation (UDA) refers to the problem of learning a model in a target domain where labeled data are not available.
This paper introduces a novel deep architecture which addresses the problem of UDA by automatically discovering latent domains in visual datasets.
We evaluate our approach on publicly available benchmarks, showing that it outperforms state-of-the-art domain adaptation methods.
arXiv Detail & Related papers (2021-03-25T14:33:33Z) - Domain Impression: A Source Data Free Domain Adaptation Method [27.19677042654432]
Unsupervised domain adaptation methods solve the adaptation problem for an unlabeled target set, assuming that the source dataset is available with all labels.
This paper proposes a domain adaptation technique that does not need any source data.
Instead of the source data, we are only provided with a classifier that is trained on the source data.
arXiv Detail & Related papers (2021-02-17T19:50:49Z) - Open-Set Hypothesis Transfer with Semantic Consistency [99.83813484934177]
We introduce a method that focuses on the semantic consistency under transformation of target data.
Our model first discovers confident predictions and performs classification with pseudo-labels.
As a result, unlabeled data can be classified into discriminative classes coincided with either source classes or unknown classes.
arXiv Detail & Related papers (2020-10-01T10:44:31Z) - Towards Fair Cross-Domain Adaptation via Generative Learning [50.76694500782927]
Domain Adaptation (DA) targets at adapting a model trained over the well-labeled source domain to the unlabeled target domain lying in different distributions.
We develop a novel Generative Few-shot Cross-domain Adaptation (GFCA) algorithm for fair cross-domain classification.
arXiv Detail & Related papers (2020-03-04T23:25:09Z) - Do We Really Need to Access the Source Data? Source Hypothesis Transfer
for Unsupervised Domain Adaptation [102.67010690592011]
Unsupervised adaptationUDA (UDA) aims to leverage the knowledge learned from a labeled source dataset to solve similar tasks in a new unlabeled domain.
Prior UDA methods typically require to access the source data when learning to adapt the model.
This work tackles a practical setting where only a trained source model is available and how we can effectively utilize such a model without source data to solve UDA problems.
arXiv Detail & Related papers (2020-02-20T03:13:58Z) - Multi-source Domain Adaptation for Visual Sentiment Classification [92.53780541232773]
We propose a novel multi-source domain adaptation (MDA) method, termed Multi-source Sentiment Generative Adversarial Network (MSGAN)
To handle data from multiple source domains, MSGAN learns to find a unified sentiment latent space where data from both the source and target domains share a similar distribution.
Extensive experiments conducted on four benchmark datasets demonstrate that MSGAN significantly outperforms the state-of-the-art MDA approaches for visual sentiment classification.
arXiv Detail & Related papers (2020-01-12T08:37:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.