Making Pre-trained Language Models End-to-end Few-shot Learners with
Contrastive Prompt Tuning
- URL: http://arxiv.org/abs/2204.00166v1
- Date: Fri, 1 Apr 2022 02:24:24 GMT
- Title: Making Pre-trained Language Models End-to-end Few-shot Learners with
Contrastive Prompt Tuning
- Authors: Ziyun Xu, Chengyu Wang, Minghui Qiu, Fuli Luo, Runxin Xu, Songfang
Huang, Jun Huang
- Abstract summary: We present CP-Tuning, the first end-to-end Contrastive Prompt Tuning framework for fine-tuning Language Models.
It is integrated with the task-invariant continuous prompt encoding technique with fully trainable prompt parameters.
Experiments over a variety of language understanding tasks used in IR systems and different PLMs show that CP-Tuning outperforms state-of-the-art methods.
- Score: 41.15017636192417
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Pre-trained Language Models (PLMs) have achieved remarkable performance for
various language understanding tasks in IR systems, which require the
fine-tuning process based on labeled training data. For low-resource scenarios,
prompt-based learning for PLMs exploits prompts as task guidance and turns
downstream tasks into masked language problems for effective few-shot
fine-tuning. In most existing approaches, the high performance of prompt-based
learning heavily relies on handcrafted prompts and verbalizers, which may limit
the application of such approaches in real-world scenarios. To solve this
issue, we present CP-Tuning, the first end-to-end Contrastive Prompt Tuning
framework for fine-tuning PLMs without any manual engineering of task-specific
prompts and verbalizers. It is integrated with the task-invariant continuous
prompt encoding technique with fully trainable prompt parameters. We further
propose the pair-wise cost-sensitive contrastive learning procedure to optimize
the model in order to achieve verbalizer-free class mapping and enhance the
task-invariance of prompts. It explicitly learns to distinguish different
classes and makes the decision boundary smoother by assigning different costs
to easy and hard cases. Experiments over a variety of language understanding
tasks used in IR systems and different PLMs show that CP-Tuning outperforms
state-of-the-art methods.
Related papers
- Parameter-Efficient Fine-Tuning of Large Language Models using Semantic Knowledge Tuning [0.08795040582681389]
Large Language Models (LLMs) are gaining significant popularity in recent years for specialized tasks using prompts.
We propose a novel method called Semantic Knowledge Tuning (SK-Tuning) for prompt and prefix tuning that employs meaningful words instead of random tokens.
Our experimental results show that SK-Tuning exhibits faster training times, fewer parameters, and superior performance on tasks such as text classification and understanding.
arXiv Detail & Related papers (2024-10-11T07:55:09Z) - OverPrompt: Enhancing ChatGPT through Efficient In-Context Learning [49.38867353135258]
We propose OverPrompt, leveraging the in-context learning capability of LLMs to handle multiple task inputs.
Our experiments show that OverPrompt can achieve cost-efficient zero-shot classification without causing significant detriment to task performance.
arXiv Detail & Related papers (2023-05-24T10:08:04Z) - Instance-wise Prompt Tuning for Pretrained Language Models [72.74916121511662]
Instance-wise Prompt Tuning (IPT) is the first prompt learning paradigm that injects knowledge from the input data instances to the prompts.
IPT significantly outperforms task-based prompt learning methods, and achieves comparable performance to conventional finetuning with only 0.5% - 1.5% of tuned parameters.
arXiv Detail & Related papers (2022-06-04T10:08:50Z) - Towards Unified Prompt Tuning for Few-shot Text Classification [47.71344780587704]
We present the Unified Prompt Tuning (UPT) framework, leading to better few-shot text classification for BERT-style models.
In UPT, a novel paradigm Prompt-Options-Verbalizer is proposed for joint prompt learning across different NLP tasks.
We also design a self-supervised task named Knowledge-enhanced Selective Masked Language Modeling to improve the PLM's generalization abilities.
arXiv Detail & Related papers (2022-05-11T07:40:45Z) - IDPG: An Instance-Dependent Prompt Generation Method [58.45110542003139]
Prompt tuning is a new, efficient NLP transfer learning paradigm that adds a task-specific prompt in each input instance during the model training stage.
We propose a conditional prompt generation method to generate prompts for each input instance.
arXiv Detail & Related papers (2022-04-09T15:45:27Z) - CINS: Comprehensive Instruction for Few-shot Learning in Task-oriented
Dialog Systems [56.302581679816775]
This paper proposes Comprehensive Instruction (CINS) that exploits PLMs with task-specific instructions.
We design a schema (definition, constraint, prompt) of instructions and their customized realizations for three important downstream tasks in ToD.
Experiments are conducted on these ToD tasks in realistic few-shot learning scenarios with small validation data.
arXiv Detail & Related papers (2021-09-10T03:23:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.