Improving the Sample Efficiency of Prompt Tuning with Domain Adaptation
- URL: http://arxiv.org/abs/2210.02952v1
- Date: Thu, 6 Oct 2022 14:44:21 GMT
- Title: Improving the Sample Efficiency of Prompt Tuning with Domain Adaptation
- Authors: Xu Guo, Boyang Li, Han Yu
- Abstract summary: We propose doMain Adaptation (OPTIMA) to regularizes the decision boundary to be smooth around regions where source and target data distributions are similar.
OPTIMA significantly enhances the transferability and sample-efficiency of prompt tuning compared to strong baselines.
- Score: 15.388175691903252
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Prompt tuning, or the conditioning of a frozen pretrained language model
(PLM) with soft prompts learned from data, has demonstrated impressive
performance on a wide range of NLP tasks. However, prompt tuning requires a
large training dataset to be effective and is outperformed by finetuning the
entire PLM in data-scarce regimes. Previous work
\citep{gu-etal-2022-ppt,vu-etal-2022-spot} proposed to transfer soft prompts
pretrained on the source domain to the target domain. In this paper, we explore
domain adaptation for prompt tuning, a problem setting where unlabeled data
from the target domain are available during pretraining. We propose bOosting
Prompt TunIng with doMain Adaptation (OPTIMA), which regularizes the decision
boundary to be smooth around regions where source and target data distributions
are similar. Extensive experiments demonstrate that OPTIMA significantly
enhances the transferability and sample-efficiency of prompt tuning compared to
strong baselines. Moreover, in few-shot settings, OPTIMA exceeds full-model
tuning by a large margin.
Related papers
- Generating Synthetic Datasets for Few-shot Prompt Tuning [48.10054761841462]
In few-shot learning settings, prompt tuning lags far behind full-model fine-tuning, limiting its scope of application.
In this paper, we leverage the powerful LLMs to synthesize task-specific labeled data for training the soft prompts.
We train soft prompts on both synthetic and real datasets using a gradient surgery approach.
arXiv Detail & Related papers (2024-10-08T01:00:02Z) - Test-Time Low Rank Adaptation via Confidence Maximization for Zero-Shot Generalization of Vision-Language Models [4.655740975414312]
This paper introduces Test-Time Low-rank adaptation (TTL) as an alternative to prompt tuning for zero-shot generalizations of large-scale vision-language models (VLMs)
TTL offers a test-time-efficient adaptation approach that updates the attention weights of the transformer by maximizing prediction confidence.
arXiv Detail & Related papers (2024-07-22T17:59:19Z) - Better Practices for Domain Adaptation [62.70267990659201]
Domain adaptation (DA) aims to provide frameworks for adapting models to deployment data without using labels.
Unclear validation protocol for DA has led to bad practices in the literature.
We show challenges across all three branches of domain adaptation methodology.
arXiv Detail & Related papers (2023-09-07T17:44:18Z) - Approximated Prompt Tuning for Vision-Language Pre-trained Models [54.326232586461614]
In vision-language pre-trained models, prompt tuning often requires a large number of learnable tokens to bridge the gap between the pre-training and downstream tasks.
We propose a novel Approximated Prompt Tuning (APT) approach towards efficient VL transfer learning.
arXiv Detail & Related papers (2023-06-27T05:43:47Z) - Automatic Data Augmentation for Domain Adapted Fine-Tuning of
Self-Supervised Speech Representations [21.423349835589793]
Self-Supervised Learning (SSL) has allowed leveraging large amounts of unlabeled speech data to improve the performance of speech recognition models.
Despite this, speech SSL representations may fail while facing an acoustic mismatch between the pretraining and target datasets.
We propose a novel supervised domain adaptation method, designed for cases exhibiting such a mismatch in acoustic domains.
arXiv Detail & Related papers (2023-06-01T09:30:49Z) - Active Finetuning: Exploiting Annotation Budget in the
Pretraining-Finetuning Paradigm [132.9949120482274]
This paper focuses on the selection of samples for annotation in the pretraining-finetuning paradigm.
We propose a novel method called ActiveFT for active finetuning task to select a subset of data distributing similarly with the entire unlabeled pool.
Extensive experiments show the leading performance and high efficiency of ActiveFT superior to baselines on both image classification and semantic segmentation.
arXiv Detail & Related papers (2023-03-25T07:17:03Z) - MADAv2: Advanced Multi-Anchor Based Active Domain Adaptation
Segmentation [98.09845149258972]
We introduce active sample selection to assist domain adaptation regarding the semantic segmentation task.
With only a little workload to manually annotate these samples, the distortion of the target-domain distribution can be effectively alleviated.
A powerful semi-supervised domain adaptation strategy is proposed to alleviate the long-tail distribution problem.
arXiv Detail & Related papers (2023-01-18T07:55:22Z) - Visual Prompt Tuning for Test-time Domain Adaptation [48.16620171809511]
We propose a simple recipe called data-efficient prompt tuning (DePT) with two key ingredients.
We find such parameter-efficient finetuning can efficiently adapt the model representation to the target domain without overfitting to the noise in the learning objective.
With much fewer parameters, DePT demonstrates not only state-of-the-art performance on major adaptation benchmarks, but also superior data efficiency.
arXiv Detail & Related papers (2022-10-10T16:45:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.