Natural Language Inference Prompts for Zero-shot Emotion Classification
in Text across Corpora
- URL: http://arxiv.org/abs/2209.06701v2
- Date: Thu, 15 Sep 2022 07:58:28 GMT
- Title: Natural Language Inference Prompts for Zero-shot Emotion Classification
in Text across Corpora
- Authors: Flor Miriam Plaza-del-Arco, Mar\'ia-Teresa Mart\'in-Valdivia, Roman
Klinger
- Abstract summary: We show that the choice of a particular prompt formulation needs to fit to the corpus.
We show that this challenge can be tackled with combinations of multiple prompts.
- Score: 11.986676998327864
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Within textual emotion classification, the set of relevant labels depends on
the domain and application scenario and might not be known at the time of model
development. This conflicts with the classical paradigm of supervised learning
in which the labels need to be predefined. A solution to obtain a model with a
flexible set of labels is to use the paradigm of zero-shot learning as a
natural language inference task, which in addition adds the advantage of not
needing any labeled training data. This raises the question how to prompt a
natural language inference model for zero-shot learning emotion classification.
Options for prompt formulations include the emotion name anger alone or the
statement "This text expresses anger". With this paper, we analyze how
sensitive a natural language inference-based zero-shot-learning classifier is
to such changes to the prompt under consideration of the corpus: How carefully
does the prompt need to be selected? We perform experiments on an established
set of emotion datasets presenting different language registers according to
different sources (tweets, events, blogs) with three natural language inference
models and show that indeed the choice of a particular prompt formulation needs
to fit to the corpus. We show that this challenge can be tackled with
combinations of multiple prompts. Such ensemble is more robust across corpora
than individual prompts and shows nearly the same performance as the individual
best prompt for a particular corpus.
Related papers
- Scene Graph Generation with Role-Playing Large Language Models [50.252588437973245]
Current approaches for open-vocabulary scene graph generation (OVSGG) use vision-language models such as CLIP.
We propose SDSGG, a scene-specific description based OVSGG framework.
To capture the complicated interplay between subjects and objects, we propose a new lightweight module called mutual visual adapter.
arXiv Detail & Related papers (2024-10-20T11:40:31Z) - Learning Phonotactics from Linguistic Informants [54.086544221761486]
Our model iteratively selects or synthesizes a data-point according to one of a range of information-theoretic policies.
We find that the information-theoretic policies that our model uses to select items to query the informant achieve sample efficiency comparable to, or greater than, fully supervised approaches.
arXiv Detail & Related papers (2024-05-08T00:18:56Z) - Paralinguistics-Enhanced Large Language Modeling of Spoken Dialogue [71.15186328127409]
Paralinguistics-enhanced Generative Pretrained Transformer (ParalinGPT)
Model takes the conversational context of text, speech embeddings, and paralinguistic attributes as input prompts within a serialized multitasking framework.
We utilize the Switchboard-1 corpus, including its sentiment labels as the paralinguistic attribute, as our spoken dialogue dataset.
arXiv Detail & Related papers (2023-12-23T18:14:56Z) - LanSER: Language-Model Supported Speech Emotion Recognition [25.597250907836152]
We present LanSER, a method that enables the use of unlabeled data by inferring weak emotion labels via pre-trained large language models.
For inferring weak labels constrained to a taxonomy, we use a textual entailment approach that selects an emotion label with the highest entailment score for a speech transcript extracted via automatic speech recognition.
Our experimental results show that models pre-trained on large datasets with this weak supervision outperform other baseline models on standard SER datasets when fine-tuned, and show improved label efficiency.
arXiv Detail & Related papers (2023-09-07T19:21:08Z) - ChatGPT as a Text Simplification Tool to Remove Bias [0.0]
The presence of specific linguistic signals particular to a certain sub-group can be picked up by language models during training.
We explore a potential technique for bias mitigation in the form of simplification of text.
arXiv Detail & Related papers (2023-05-09T13:10:23Z) - LabelPrompt: Effective Prompt-based Learning for Relation Classification [31.291466190218912]
This paper presents a novel prompt-based learning method, namely LabelPrompt, for the relation classification task.
Motivated by the intuition to GIVE MODEL CHOICES!'', we first define additional tokens to represent relation labels, which regard these tokens as the verbaliser with semantic initialisation.
Then, to mitigate inconsistency between predicted relations and given entities, we implement an entity-aware module with contrastive learning.
arXiv Detail & Related papers (2023-02-16T04:06:25Z) - CCPrefix: Counterfactual Contrastive Prefix-Tuning for Many-Class
Classification [57.62886091828512]
We propose a brand-new prefix-tuning method, Counterfactual Contrastive Prefix-tuning (CCPrefix) for many-class classification.
Basically, an instance-dependent soft prefix, derived from fact-counterfactual pairs in the label space, is leveraged to complement the language verbalizers in many-class classification.
arXiv Detail & Related papers (2022-11-11T03:45:59Z) - Discriminative Language Model as Semantic Consistency Scorer for
Prompt-based Few-Shot Text Classification [10.685862129925727]
This paper proposes a novel prompt-based finetuning method (called DLM-SCS) for few-shot text classification.
The underlying idea is that the prompt instantiated with the true label should have higher semantic consistency score than other prompts with false labels.
Our model outperforms several state-of-the-art prompt-based few-shot methods.
arXiv Detail & Related papers (2022-10-23T16:10:48Z) - Eliciting Knowledge from Pretrained Language Models for Prototypical
Prompt Verbalizer [12.596033546002321]
In this paper, we focus on eliciting knowledge from pretrained language models and propose a prototypical prompt verbalizer for prompt-tuning.
For zero-shot settings, knowledge is elicited from pretrained language models by a manually designed template to form initial prototypical embeddings.
For few-shot settings, models are tuned to learn meaningful and interpretable prototypical embeddings.
arXiv Detail & Related papers (2022-01-14T12:04:37Z) - On The Ingredients of an Effective Zero-shot Semantic Parser [95.01623036661468]
We analyze zero-shot learning by paraphrasing training examples of canonical utterances and programs from a grammar.
We propose bridging these gaps using improved grammars, stronger paraphrasers, and efficient learning methods.
Our model achieves strong performance on two semantic parsing benchmarks (Scholar, Geo) with zero labeled data.
arXiv Detail & Related papers (2021-10-15T21:41:16Z) - Leveraging Adversarial Training in Self-Learning for Cross-Lingual Text
Classification [52.69730591919885]
We present a semi-supervised adversarial training process that minimizes the maximal loss for label-preserving input perturbations.
We observe significant gains in effectiveness on document and intent classification for a diverse set of languages.
arXiv Detail & Related papers (2020-07-29T19:38:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.