Reframing Instructional Prompts to GPTk's Language
- URL: http://arxiv.org/abs/2109.07830v1
- Date: Thu, 16 Sep 2021 09:44:43 GMT
- Title: Reframing Instructional Prompts to GPTk's Language
- Authors: Swaroop Mishra, Daniel Khashabi, Chitta Baral, Yejin Choi and Hannaneh
Hajishirzi
- Abstract summary: We propose reframing techniques for model designers to create effective prompts for language models.
Our results show that reframing improves few-shot learning performance by 14% while reducing sample complexity.
The performance gains are particularly important on large language models, such as GPT3 where tuning models or prompts on large datasets is not feasible.
- Score: 72.69833640335519
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: How can model designers turn task instructions into effective prompts for
language models? Backed by extensive empirical analysis on GPT3, we observe
important features for successful instructional prompts, and propose several
reframing techniques for model designers to create such prompts. For example, a
complex task can be decomposed into multiple simpler tasks. We experiment over
12 NLP tasks across 6 diverse categories (question generation, classification,
etc.). Our results show that reframing improves few-shot learning performance
by 14\% while reducing sample complexity over existing few-shot baselines. The
performance gains are particularly important on large language models, such as
GPT3 where tuning models or prompts on large datasets is not feasible.
Furthermore, we observe that such gains are not limited to GPT3; the reframed
tasks remain superior over raw instructions across different model
architectures, underscoring the cross-model generality of these guidelines. We
hope these empirical-driven techniques will pave way for more effective ways to
prompt LMs in future.
Related papers
- Deconstructing In-Context Learning: Understanding Prompts via Corruption [13.37109575313212]
We decompose the entire prompt into four components: task description, demonstration inputs, labels, and inline instructions.
We study models ranging from 1.5B to 70B in size, using ten datasets covering classification and generation tasks.
We find that repeating text within the prompt boosts model performance, and bigger models are more sensitive to the semantics of the prompt.
arXiv Detail & Related papers (2024-04-02T15:50:55Z) - Rethinking ASTE: A Minimalist Tagging Scheme Alongside Contrastive Learning [7.785948823258398]
Aspect Sentiment Triplet Extraction (ASTE) is a burgeoning subtask of fine-grained sentiment analysis.
Existing approaches to ASTE often complicate the task with additional structures or external data.
We propose a novel tagging scheme and employ a contrastive learning approach to mitigate these challenges.
arXiv Detail & Related papers (2024-03-12T06:01:04Z) - Mixture-of-Experts Meets Instruction Tuning:A Winning Combination for
Large Language Models [125.91897197446379]
We find that MoE models benefit more from instruction tuning than dense models.
Our most powerful model, FLAN-MOE-32B, surpasses the performance of FLAN-PALM-62B on four benchmark tasks.
arXiv Detail & Related papers (2023-05-24T04:22:26Z) - Instruction Induction: From Few Examples to Natural Language Task
Descriptions [55.139554327372934]
We show that language models can explicitly infer an underlying task from a few demonstrations by prompting them to generate a natural language instruction that fits the examples.
InstructGPT achieves 65.7% of human performance in our execution-based metric, while the original GPT-3 model reaches only 9.8% of human performance.
arXiv Detail & Related papers (2022-05-22T09:22:37Z) - CINS: Comprehensive Instruction for Few-shot Learning in Task-oriented
Dialog Systems [56.302581679816775]
This paper proposes Comprehensive Instruction (CINS) that exploits PLMs with task-specific instructions.
We design a schema (definition, constraint, prompt) of instructions and their customized realizations for three important downstream tasks in ToD.
Experiments are conducted on these ToD tasks in realistic few-shot learning scenarios with small validation data.
arXiv Detail & Related papers (2021-09-10T03:23:06Z) - Finetuned Language Models Are Zero-Shot Learners [67.70352207685558]
We show that instruction tuning boosts zero-shot performance on unseen tasks.
We take a 137B parameter pretrained language model and instruction-tune it on over 60 NLP tasks verbalized via natural language instruction templates.
We evaluate this instruction-tuned model, which we call FLAN, on unseen task types.
arXiv Detail & Related papers (2021-09-03T17:55:52Z) - Language Models are Few-Shot Learners [61.36677350504291]
We show that scaling up language models greatly improves task-agnostic, few-shot performance.
We train GPT-3, an autoregressive language model with 175 billion parameters, and test its performance in the few-shot setting.
GPT-3 achieves strong performance on many NLP datasets, including translation, question-answering, and cloze tasks.
arXiv Detail & Related papers (2020-05-28T17:29:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.