PADA: A Prompt-based Autoregressive Approach for Adaptation to Unseen
Domains
- URL: http://arxiv.org/abs/2102.12206v1
- Date: Wed, 24 Feb 2021 11:02:29 GMT
- Title: PADA: A Prompt-based Autoregressive Approach for Adaptation to Unseen
Domains
- Authors: Eyal Ben-David, Nadav Oved, Roi Reichart
- Abstract summary: PADA: A Prompt-based Autoregressive Domain Adaptation algorithm, based on the T5 model.
We present PADA: A Prompt-based Autoregressive Domain Adaptation algorithm, based on the T5 model.
In experiments with two tasks, PADA strongly outperforms state-of-the-art approaches and additional strong baselines.
- Score: 19.682729518136142
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Natural Language Processing algorithms have made incredible progress
recently, but they still struggle when applied to out-of-distribution examples.
In this paper, we address a very challenging and previously underexplored
version of this domain adaptation problem. In our setup an algorithm is trained
on several source domains, and then applied to examples from an unseen domain
that is unknown at training time. Particularly, no examples, labeled or
unlabeled, or any other knowledge about the target domain are available to the
algorithm at training time. We present PADA: A Prompt-based Autoregressive
Domain Adaptation algorithm, based on the T5 model. Given a test example, PADA
first generates a unique prompt and then, conditioned on this prompt, labels
the example with respect to the NLP task. The prompt is a sequence of
unrestricted length, consisting of pre-defined Domain Related Features (DRFs)
that characterize each of the source domains. Intuitively, the prompt is a
unique signature that maps the test example to the semantic space spanned by
the source domains. In experiments with two tasks: Rumour Detection and
Multi-Genre Natural Language Inference (MNLI), for a total of 10 multi-source
adaptation scenarios, PADA strongly outperforms state-of-the-art approaches and
additional strong baselines.
Related papers
- Domain-Controlled Prompt Learning [49.45309818782329]
Existing prompt learning methods often lack domain-awareness or domain-transfer mechanisms.
We propose a textbfDomain-Controlled Prompt Learning for the specific domains.
Our method achieves state-of-the-art performance in specific domain image recognition datasets.
arXiv Detail & Related papers (2023-09-30T02:59:49Z) - SwitchPrompt: Learning Domain-Specific Gated Soft Prompts for
Classification in Low-Resource Domains [14.096170976149521]
SwitchPrompt is a novel and lightweight prompting methodology for adaptation of language models trained on datasets from the general domain to diverse low-resource domains.
Our few-shot experiments on three text classification benchmarks demonstrate the efficacy of the general-domain pre-trained language models when used with SwitchPrompt.
They often even outperform their domain-specific counterparts trained with baseline state-of-the-art prompting methods by up to 10.7% performance increase in accuracy.
arXiv Detail & Related papers (2023-02-14T07:14:08Z) - Multi-Prompt Alignment for Multi-Source Unsupervised Domain Adaptation [86.02485817444216]
We introduce Multi-Prompt Alignment (MPA), a simple yet efficient framework for multi-source UDA.
MPA denoises the learned prompts through an auto-encoding process and aligns them by maximizing the agreement of all the reconstructed prompts.
Experiments show that MPA achieves state-of-the-art results on three popular datasets with an impressive average accuracy of 54.1% on DomainNet.
arXiv Detail & Related papers (2022-09-30T03:40:10Z) - Example-based Hypernetworks for Out-of-Distribution Generalization [43.99982893976026]
This paper addresses the issue of multi-source adaptation for unfamiliar domains.
We leverage labeled data from multiple source domains to generalize to unknown target domains at training.
Our innovative framework employs example-based Hypernetwork adaptation.
arXiv Detail & Related papers (2022-03-27T11:10:10Z) - Dynamic Instance Domain Adaptation [109.53575039217094]
Most studies on unsupervised domain adaptation assume that each domain's training samples come with domain labels.
We develop a dynamic neural network with adaptive convolutional kernels to generate instance-adaptive residuals to adapt domain-agnostic deep features to each individual instance.
Our model, dubbed DIDA-Net, achieves state-of-the-art performance on several commonly used single-source and multi-source UDA datasets.
arXiv Detail & Related papers (2022-03-09T20:05:54Z) - DoCoGen: Domain Counterfactual Generation for Low Resource Domain
Adaptation [19.834322227607917]
We propose a controllable generation approach to deal with domain adaptation challenges.
Our model outperforms strong baselines and improves the accuracy of a state-of-the-art unsupervised DA algorithm.
arXiv Detail & Related papers (2022-02-24T20:25:46Z) - Domain Adaptation via Prompt Learning [39.97105851723885]
Unsupervised domain adaption (UDA) aims to adapt models learned from a well-annotated source domain to a target domain.
We introduce a novel prompt learning paradigm for UDA, named Domain Adaptation via Prompt Learning (DAPL)
arXiv Detail & Related papers (2022-02-14T13:25:46Z) - Batch Normalization Embeddings for Deep Domain Generalization [50.51405390150066]
Domain generalization aims at training machine learning models to perform robustly across different and unseen domains.
We show a significant increase in classification accuracy over current state-of-the-art techniques on popular domain generalization benchmarks.
arXiv Detail & Related papers (2020-11-25T12:02:57Z) - CMT in TREC-COVID Round 2: Mitigating the Generalization Gaps from Web
to Special Domain Search [89.48123965553098]
This paper presents a search system to alleviate the special domain adaption problem.
The system utilizes the domain-adaptive pretraining and few-shot learning technologies to help neural rankers mitigate the domain discrepancy.
Our system performs the best among the non-manual runs in Round 2 of the TREC-COVID task.
arXiv Detail & Related papers (2020-11-03T09:10:48Z) - Text Recognition in Real Scenarios with a Few Labeled Samples [55.07859517380136]
Scene text recognition (STR) is still a hot research topic in computer vision field.
This paper proposes a few-shot adversarial sequence domain adaptation (FASDA) approach to build sequence adaptation.
Our approach can maximize the character-level confusion between the source domain and the target domain.
arXiv Detail & Related papers (2020-06-22T13:03:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.