CCPrefix: Counterfactual Contrastive Prefix-Tuning for Many-Class
Classification
- URL: http://arxiv.org/abs/2211.05987v2
- Date: Tue, 13 Feb 2024 02:51:03 GMT
- Title: CCPrefix: Counterfactual Contrastive Prefix-Tuning for Many-Class
Classification
- Authors: Yang Li, Canran Xu, Guodong Long, Tao Shen, Chongyang Tao and Jing
Jiang
- Abstract summary: We propose a brand-new prefix-tuning method, Counterfactual Contrastive Prefix-tuning (CCPrefix) for many-class classification.
Basically, an instance-dependent soft prefix, derived from fact-counterfactual pairs in the label space, is leveraged to complement the language verbalizers in many-class classification.
- Score: 57.62886091828512
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Recently, prefix-tuning was proposed to efficiently adapt pre-trained
language models to a broad spectrum of natural language classification tasks.
It leverages soft prefix as task-specific indicators and language verbalizers
as categorical-label mentions to narrow the formulation gap from pre-training
language models. However, when the label space increases considerably (i.e.,
many-class classification), such a tuning technique suffers from a verbalizer
ambiguity problem since the many-class labels are represented by
semantic-similar verbalizers in short language phrases. To overcome this,
inspired by the human-decision process that the most ambiguous classes would be
mulled over for each instance, we propose a brand-new prefix-tuning method,
Counterfactual Contrastive Prefix-tuning (CCPrefix), for many-class
classification. Basically, an instance-dependent soft prefix, derived from
fact-counterfactual pairs in the label space, is leveraged to complement the
language verbalizers in many-class classification. We conduct experiments on
many-class benchmark datasets in both the fully supervised setting and the
few-shot setting, which indicates that our model outperforms former baselines.
Related papers
- Manual Verbalizer Enrichment for Few-Shot Text Classification [1.860409237919611]
acrshortmave is an approach for verbalizer construction by enrichment of class labels.
Our model achieves state-of-the-art results while using significantly fewer resources.
arXiv Detail & Related papers (2024-10-08T16:16:47Z) - SciPrompt: Knowledge-augmented Prompting for Fine-grained Categorization of Scientific Topics [2.3742710594744105]
We introduce SciPrompt, a framework designed to automatically retrieve scientific topic-related terms for low-resource text classification tasks.
Our method outperforms state-of-the-art, prompt-based fine-tuning methods on scientific text classification tasks under few and zero-shot settings.
arXiv Detail & Related papers (2024-10-02T18:45:04Z) - Language Models for Text Classification: Is In-Context Learning Enough? [54.869097980761595]
Recent foundational language models have shown state-of-the-art performance in many NLP tasks in zero- and few-shot settings.
An advantage of these models over more standard approaches is the ability to understand instructions written in natural language (prompts)
This makes them suitable for addressing text classification problems for domains with limited amounts of annotated instances.
arXiv Detail & Related papers (2024-03-26T12:47:39Z) - A Novel Prompt-tuning Method: Incorporating Scenario-specific Concepts
into a Verbalizer [15.612761980503658]
We propose a label-word construction process that incorporates scenario-specific concepts.
Specifically, we extract rich concepts from task-specific scenarios as label-word candidates.
We develop a novel cascade calibration module to refine the candidates into a set of label words for each class.
arXiv Detail & Related papers (2024-01-10T15:02:35Z) - Mitigating Word Bias in Zero-shot Prompt-based Classifiers [55.60306377044225]
We show that matching class priors correlates strongly with the oracle upper bound performance.
We also demonstrate large consistent performance gains for prompt settings over a range of NLP tasks.
arXiv Detail & Related papers (2023-09-10T10:57:41Z) - On Guiding Visual Attention with Language Specification [76.08326100891571]
We use high-level language specification as advice for constraining the classification evidence to task-relevant features, instead of distractors.
We show that supervising spatial attention in this way improves performance on classification tasks with biased and noisy data.
arXiv Detail & Related papers (2022-02-17T22:40:19Z) - Eliciting Knowledge from Pretrained Language Models for Prototypical
Prompt Verbalizer [12.596033546002321]
In this paper, we focus on eliciting knowledge from pretrained language models and propose a prototypical prompt verbalizer for prompt-tuning.
For zero-shot settings, knowledge is elicited from pretrained language models by a manually designed template to form initial prototypical embeddings.
For few-shot settings, models are tuned to learn meaningful and interpretable prototypical embeddings.
arXiv Detail & Related papers (2022-01-14T12:04:37Z) - Knowledgeable Prompt-tuning: Incorporating Knowledge into Prompt
Verbalizer for Text Classification [68.3291372168167]
We focus on incorporating external knowledge into the verbalizer, forming a knowledgeable prompt-tuning (KPT)
We expand the label word space of the verbalizer using external knowledge bases (KBs) and refine the expanded label word space with the PLM itself before predicting with the expanded label word space.
Experiments on zero and few-shot text classification tasks demonstrate the effectiveness of knowledgeable prompt-tuning.
arXiv Detail & Related papers (2021-08-04T13:00:16Z) - Leveraging Adversarial Training in Self-Learning for Cross-Lingual Text
Classification [52.69730591919885]
We present a semi-supervised adversarial training process that minimizes the maximal loss for label-preserving input perturbations.
We observe significant gains in effectiveness on document and intent classification for a diverse set of languages.
arXiv Detail & Related papers (2020-07-29T19:38:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.