Toward Human Readable Prompt Tuning: Kubrick's The Shining is a good
movie, and a good prompt too?
- URL: http://arxiv.org/abs/2212.10539v1
- Date: Tue, 20 Dec 2022 18:47:13 GMT
- Title: Toward Human Readable Prompt Tuning: Kubrick's The Shining is a good
movie, and a good prompt too?
- Authors: Weijia Shi, Xiaochuang Han, Hila Gonen, Ari Holtzman, Yulia Tsvetkov,
Luke Zettlemoyer
- Abstract summary: Large language models can perform new tasks in a zero-shot fashion, given natural language prompts.
It is underexplored what factors make the prompts effective, especially when the prompts are natural language.
- Score: 84.91689960190054
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Large language models can perform new tasks in a zero-shot fashion, given
natural language prompts that specify the desired behavior. Such prompts are
typically hand engineered, but can also be learned with gradient-based methods
from labeled data. However, it is underexplored what factors make the prompts
effective, especially when the prompts are natural language. In this paper, we
investigate common attributes shared by effective prompts. We first propose a
human readable prompt tuning method (F LUENT P ROMPT) based on Langevin
dynamics that incorporates a fluency constraint to find a diverse distribution
of effective and fluent prompts. Our analysis reveals that effective prompts
are topically related to the task domain and calibrate the prior probability of
label words. Based on these findings, we also propose a method for generating
prompts using only unlabeled data, outperforming strong baselines by an average
of 7.0% accuracy across three tasks.
Related papers
- Prompt Valuation Based on Shapley Values [5.072508764734943]
Large language models (LLMs) excel on new tasks without additional training, simply by providing natural language prompts.
In this paper, we utilize the Shapley value to fairly quantify the contributions of prompts.
We validate the effectiveness of using the Shapley value for prompts as it effectively distinguishes and quantifies the contributions of each prompt.
arXiv Detail & Related papers (2023-12-24T03:37:11Z) - The language of prompting: What linguistic properties make a prompt
successful? [13.034603322224548]
LLMs can be prompted to achieve impressive zero-shot or few-shot performance in many NLP tasks.
Yet, we still lack a systematic understanding of how linguistic properties of prompts correlate with task performance.
We investigate both grammatical properties such as mood, tense, aspect and modality, as well as lexico-semantic variation through the use of synonyms.
arXiv Detail & Related papers (2023-11-03T15:03:36Z) - InfoPrompt: Information-Theoretic Soft Prompt Tuning for Natural
Language Understanding [51.48361798508375]
We develop an information-theoretic framework that formulates soft prompt tuning as maximizing mutual information between prompts and other model parameters.
We show that InfoPrompt can significantly accelerate the convergence of the prompt tuning and outperform traditional prompt tuning methods.
arXiv Detail & Related papers (2023-06-08T04:31:48Z) - Zero-shot Approach to Overcome Perturbation Sensitivity of Prompts [7.208567411886273]
Recent studies have demonstrated that natural-language prompts can help to leverage the knowledge learned by pre-trained language models for the binary sentence-level sentiment classification task.
This study aims to find high-quality prompts for the given task in a zero-shot setting.
We empirically demonstrate that the top-ranked prompts are high-quality and significantly outperform the base prompt and the prompts generated using few-shot learning for the binary sentence-level sentiment classification task.
arXiv Detail & Related papers (2023-05-25T03:36:43Z) - Demystifying Prompts in Language Models via Perplexity Estimation [109.59105230163041]
Performance of a prompt is coupled with the extent to which the model is familiar with the language it contains.
We show that the lower the perplexity of the prompt is, the better the prompt is able to perform the task.
arXiv Detail & Related papers (2022-12-08T02:21:47Z) - TEMPERA: Test-Time Prompting via Reinforcement Learning [57.48657629588436]
We propose Test-time Prompt Editing using Reinforcement learning (TEMPERA)
In contrast to prior prompt generation methods, TEMPERA can efficiently leverage prior knowledge.
Our method achieves 5.33x on average improvement in sample efficiency when compared to the traditional fine-tuning methods.
arXiv Detail & Related papers (2022-11-21T22:38:20Z) - Bayesian Prompt Learning for Image-Language Model Generalization [64.50204877434878]
We use the regularization ability of Bayesian methods to frame prompt learning as a variational inference problem.
Our approach regularizes the prompt space, reduces overfitting to the seen prompts and improves the prompt generalization on unseen prompts.
We demonstrate empirically on 15 benchmarks that Bayesian prompt learning provides an appropriate coverage of the prompt space.
arXiv Detail & Related papers (2022-10-05T17:05:56Z) - RLPrompt: Optimizing Discrete Text Prompts With Reinforcement Learning [84.75064077323098]
This paper proposes RLPrompt, an efficient discrete prompt optimization approach with reinforcement learning (RL)
RLPrompt is flexibly applicable to different types of LMs, such as masked gibberish (e.g., grammaBERT) and left-to-right models (e.g., GPTs)
Experiments on few-shot classification and unsupervised text style transfer show superior performance over a wide range of existing finetuning or prompting methods.
arXiv Detail & Related papers (2022-05-25T07:50:31Z) - Do Prompts Solve NLP Tasks Using Natural Language? [18.611748762251494]
In this work, we empirically compare the three types of prompts under both few-shot and fully-supervised settings.
Our experimental results show that schema prompts are the most effective in general.
arXiv Detail & Related papers (2022-03-02T07:20:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.