Structured Prompt Tuning
- URL: http://arxiv.org/abs/2205.12309v1
- Date: Tue, 24 May 2022 18:36:34 GMT
- Title: Structured Prompt Tuning
- Authors: Chi-Liang Liu, Hung-yi Lee, Wen-tau Yih
- Abstract summary: Instead of prepending a sequence of tunable embeddings to the input, we generate the soft prompt embeddings through a hypernetwork.
Our approach subsumes the standard prompt tuning, allows more flexibility in model design and can be applied to both single-task and multi-task training settings.
- Score: 83.71253868369999
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We propose structured prompt tuning, a simple and effective method to improve
prompt tuning. Instead of prepending a sequence of tunable embeddings to the
input, we generate the soft prompt embeddings through a hypernetwork. Our
approach subsumes the standard prompt tuning, allows more flexibility in model
design and can be applied to both single-task and multi-task training settings.
Empirically, structured prompt tuning shows a gain of +1.2$~1.5 points on the
GLUE benchmark and is less sensitive to the change of learning rate, compared
to standard prompt tuning.
Related papers
- Improving Prompt Tuning with Learned Prompting Layers [12.46460062708119]
We propose a novel framework, underlineSelective underlinePrompt underlineTuning (SPT)
It learns to select the proper prompt layers by inserting a prompt controlled by a learnable probabilistic gate at each intermediate layer.
We conduct extensive experiments with ten benchmark datasets under the full-data and few-shot scenarios.
arXiv Detail & Related papers (2023-10-31T02:07:51Z) - InfoPrompt: Information-Theoretic Soft Prompt Tuning for Natural
Language Understanding [51.48361798508375]
We develop an information-theoretic framework that formulates soft prompt tuning as maximizing mutual information between prompts and other model parameters.
We show that InfoPrompt can significantly accelerate the convergence of the prompt tuning and outperform traditional prompt tuning methods.
arXiv Detail & Related papers (2023-06-08T04:31:48Z) - Residual Prompt Tuning: Improving Prompt Tuning with Residual
Reparameterization [57.379285443780894]
Residual Prompt Tuning is a simple and efficient method that significantly improves the performance and stability of prompt tuning.
We show that our method reaches +7 points improvement over prompt tuning with T5-Base and allows to reduce the prompt length by 10x without hurting performance.
arXiv Detail & Related papers (2023-05-06T05:35:14Z) - PTP: Boosting Stability and Performance of Prompt Tuning with
Perturbation-Based Regularizer [94.23904400441957]
We introduce perturbation-based regularizers, which can smooth the loss landscape, into prompt tuning.
We design two kinds of perturbation-based regularizers, including random-noise-based and adversarial-based.
Our new algorithms improve the state-of-the-art prompt tuning methods by 1.94% and 2.34% on SuperGLUE and FewGLUE benchmarks, respectively.
arXiv Detail & Related papers (2023-05-03T20:30:51Z) - Multitask Pre-training of Modular Prompt for Chinese Few-Shot Learning [83.10861551885321]
We present Multi-task Pre-trained Modular Prompt (MP2) to boost prompt tuning for few-shot learning.
MP2 is a set of combinable prompts pre-trained on 38 Chinese tasks.
We show MP2 significantly outperforms prompt tuning, full model tuning, and prior prompt pre-training methods in few-shot settings.
arXiv Detail & Related papers (2022-10-14T06:43:42Z) - Prompt Tuning for Generative Multimodal Pretrained Models [75.44457974275154]
We implement prompt tuning on the unified sequence-to-sequence pretrained model adaptive to both understanding and generation tasks.
Experimental results demonstrate that the light-weight prompt tuning can achieve comparable performance with finetuning.
In comparison with finetuned models, the prompt-tuned models demonstrate improved robustness against adversarial attacks.
arXiv Detail & Related papers (2022-08-04T08:56:38Z) - PPT: Pre-trained Prompt Tuning for Few-shot Learning [47.05554619258627]
Prompts for pre-trained language models (PLMs) have shown remarkable performance by bridging the gap between pre-training tasks and various downstream tasks.
Among these methods, prompt tuning, which freezes PLMs and only tunes soft prompts, provides an efficient and effective solution for adapting large-scale PLMs to downstream tasks.
In our work, we find that prompt tuning performs comparably with conventional full-model fine-tuning when downstream data are sufficient, whereas it performs much worse under few-shot learning settings.
arXiv Detail & Related papers (2021-09-09T15:11:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.