Improving Task Generalization via Unified Schema Prompt
- URL: http://arxiv.org/abs/2208.03229v1
- Date: Fri, 5 Aug 2022 15:26:36 GMT
- Title: Improving Task Generalization via Unified Schema Prompt
- Authors: Wanjun Zhong, Yifan Gao, Ning Ding, Zhiyuan Liu, Ming Zhou, Jiahai
Wang, Jian Yin, Nan Duan
- Abstract summary: Unified Prompt is a flexible and prompting method, which automatically customizes the learnable prompts for each task according to the task input schema.
It models the shared knowledge between tasks, while keeping the characteristics of different task schema.
The framework achieves strong zero-shot and few-shot performance on 16 unseen tasks downstream from 8 task types.
- Score: 87.31158568180514
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Task generalization has been a long standing challenge in Natural Language
Processing (NLP). Recent research attempts to improve the task generalization
ability of pre-trained language models by mapping NLP tasks into human-readable
prompted forms. However, these approaches require laborious and inflexible
manual collection of prompts, and different prompts on the same downstream task
may receive unstable performance. We propose Unified Schema Prompt, a flexible
and extensible prompting method, which automatically customizes the learnable
prompts for each task according to the task input schema. It models the shared
knowledge between tasks, while keeping the characteristics of different task
schema, and thus enhances task generalization ability. The schema prompt takes
the explicit data structure of each task to formulate prompts so that little
human effort is involved. To test the task generalization ability of schema
prompt at scale, we conduct schema prompt-based multitask pre-training on a
wide variety of general NLP tasks. The framework achieves strong zero-shot and
few-shot generalization performance on 16 unseen downstream tasks from 8 task
types (e.g., QA, NLI, etc). Furthermore, comprehensive analyses demonstrate the
effectiveness of each component in the schema prompt, its flexibility in task
compositionality, and its ability to improve performance under a full-data
fine-tuning setting.
Related papers
- Active Instruction Tuning: Improving Cross-Task Generalization by
Training on Prompt Sensitive Tasks [101.40633115037983]
Instruction tuning (IT) achieves impressive zero-shot generalization results by training large language models (LLMs) on a massive amount of diverse tasks with instructions.
How to select new tasks to improve the performance and generalizability of IT models remains an open question.
We propose active instruction tuning based on prompt uncertainty, a novel framework to identify informative tasks, and then actively tune the models on the selected tasks.
arXiv Detail & Related papers (2023-11-01T04:40:05Z) - UniverSLU: Universal Spoken Language Understanding for Diverse Tasks with Natural Language Instructions [64.50935101415776]
We build a single model that jointly performs various spoken language understanding (SLU) tasks.
We demonstrate the efficacy of our single multi-task learning model "UniverSLU" for 12 speech classification and sequence generation task types spanning 17 datasets and 9 languages.
arXiv Detail & Related papers (2023-10-04T17:10:23Z) - Musketeer: Joint Training for Multi-task Vision Language Model with Task Explanation Prompts [75.75548749888029]
We present a vision-language model whose parameters are jointly trained on all tasks and fully shared among multiple heterogeneous tasks.
With a single model, Musketeer achieves results comparable to or better than strong baselines trained on single tasks, almost uniformly across multiple tasks.
arXiv Detail & Related papers (2023-05-11T17:57:49Z) - Decomposed Prompting: A Modular Approach for Solving Complex Tasks [55.42850359286304]
We propose Decomposed Prompting to solve complex tasks by decomposing them (via prompting) into simpler sub-tasks.
This modular structure allows each prompt to be optimized for its specific sub-task.
We show that the flexibility and modularity of Decomposed Prompting allows it to outperform prior work on few-shot prompting.
arXiv Detail & Related papers (2022-10-05T17:28:20Z) - Prompt Tuning with Soft Context Sharing for Vision-Language Models [42.61889428498378]
We propose a novel method to tune pre-trained vision-language models on multiple target few-shot tasks jointly.
We show that SoftCPT significantly outperforms single-task prompt tuning methods.
arXiv Detail & Related papers (2022-08-29T10:19:10Z) - Towards Unified Prompt Tuning for Few-shot Text Classification [47.71344780587704]
We present the Unified Prompt Tuning (UPT) framework, leading to better few-shot text classification for BERT-style models.
In UPT, a novel paradigm Prompt-Options-Verbalizer is proposed for joint prompt learning across different NLP tasks.
We also design a self-supervised task named Knowledge-enhanced Selective Masked Language Modeling to improve the PLM's generalization abilities.
arXiv Detail & Related papers (2022-05-11T07:40:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.