Gen-Z: Generative Zero-Shot Text Classification with Contextualized
Label Descriptions
- URL: http://arxiv.org/abs/2311.07115v1
- Date: Mon, 13 Nov 2023 07:12:57 GMT
- Title: Gen-Z: Generative Zero-Shot Text Classification with Contextualized
Label Descriptions
- Authors: Sachin Kumar, Chan Young Park, Yulia Tsvetkov
- Abstract summary: We propose a generative prompting framework for zero-shot text classification.
GEN-Z measures the LM likelihood of input text conditioned on natural language descriptions of labels.
We show that zero-shot classification with simple contextualization of the data source consistently outperforms both zero-shot and few-shot baselines.
- Score: 50.92702206798324
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Language model (LM) prompting--a popular paradigm for solving NLP tasks--has
been shown to be susceptible to miscalibration and brittleness to slight prompt
variations, caused by its discriminative prompting approach, i.e., predicting
the label given the input. To address these issues, we propose Gen-Z--a
generative prompting framework for zero-shot text classification. GEN-Z is
generative, as it measures the LM likelihood of input text, conditioned on
natural language descriptions of labels. The framework is multivariate, as
label descriptions allow us to seamlessly integrate additional contextual
information about the labels to improve task performance. On various standard
classification benchmarks, with six open-source LM families, we show that
zero-shot classification with simple contextualization of the data source of
the evaluation set consistently outperforms both zero-shot and few-shot
baselines while improving robustness to prompt variations. Further, our
approach enables personalizing classification in a zero-shot manner by
incorporating author, subject, or reader information in the label descriptions.
Related papers
- CoA: Chain-of-Action for Generative Semantic Labels [5.016605351534376]
Chain-of-Action (CoA) method generates labels aligned with contextually relevant features of an image.
CoA is designed based on the observation that enriched and valuable contextual information improves generative performance during inference.
arXiv Detail & Related papers (2024-11-26T13:09:14Z) - Description-Enhanced Label Embedding Contrastive Learning for Text
Classification [65.01077813330559]
Self-Supervised Learning (SSL) in model learning process and design a novel self-supervised Relation of Relation (R2) classification task.
Relation of Relation Learning Network (R2-Net) for text classification, in which text classification and R2 classification are treated as optimization targets.
external knowledge from WordNet to obtain multi-aspect descriptions for label semantic learning.
arXiv Detail & Related papers (2023-06-15T02:19:34Z) - Label Agnostic Pre-training for Zero-shot Text Classification [4.9081735096855565]
In real-world applications, there exists an infinite label space for describing a given text.
We introduce two new simple yet effective pre-training strategies, Implicit and Explicit pre-training.
These methods inject aspect-level understanding into the model at train time with the goal of conditioning the model to build task-level understanding.
arXiv Detail & Related papers (2023-05-25T22:55:32Z) - Imprecise Label Learning: A Unified Framework for Learning with Various Imprecise Label Configurations [91.67511167969934]
imprecise label learning (ILL) is a framework for the unification of learning with various imprecise label configurations.
We demonstrate that ILL can seamlessly adapt to partial label learning, semi-supervised learning, noisy label learning, and, more importantly, a mixture of these settings.
arXiv Detail & Related papers (2023-05-22T04:50:28Z) - The Benefits of Label-Description Training for Zero-Shot Text
Classification [35.27224341685012]
Pretrained language models have improved zero-shot text classification.
We propose a simple way to further improve zero-shot accuracies with minimal effort.
arXiv Detail & Related papers (2023-05-03T16:19:31Z) - Empowering Sentence Encoders with Prompting and Label Retrieval for
Zero-shot Text Classification [5.484132137132862]
Our framework, RaLP, encodes prompted label candidates with a sentence encoder, then assigns the label whose prompt embedding has the highest similarity with the input text embedding.
RaLP achieves competitive or stronger performance than much larger baselines on various closed-set classification and multiple-choice QA datasets.
arXiv Detail & Related papers (2022-12-20T16:18:03Z) - Self-Adaptive Label Augmentation for Semi-supervised Few-shot
Classification [121.63992191386502]
Few-shot classification aims to learn a model that can generalize well to new tasks when only a few labeled samples are available.
We propose a semi-supervised few-shot classification method that assigns an appropriate label to each unlabeled sample by a manually defined metric.
A major novelty of SALA is the task-adaptive metric, which can learn the metric adaptively for different tasks in an end-to-end fashion.
arXiv Detail & Related papers (2022-06-16T13:14:03Z) - Unsupervised Ranking and Aggregation of Label Descriptions for Zero-Shot
Classifiers [8.434227773463022]
In a true zero-shot setup, designing good label descriptions is challenging because no development set is available.
We look at how probabilistic models of repeated rating analysis can be used for selecting the best label descriptions in an unsupervised fashion.
arXiv Detail & Related papers (2022-04-20T14:23:09Z) - Knowledgeable Prompt-tuning: Incorporating Knowledge into Prompt
Verbalizer for Text Classification [68.3291372168167]
We focus on incorporating external knowledge into the verbalizer, forming a knowledgeable prompt-tuning (KPT)
We expand the label word space of the verbalizer using external knowledge bases (KBs) and refine the expanded label word space with the PLM itself before predicting with the expanded label word space.
Experiments on zero and few-shot text classification tasks demonstrate the effectiveness of knowledgeable prompt-tuning.
arXiv Detail & Related papers (2021-08-04T13:00:16Z) - Unsupervised Person Re-identification via Multi-label Classification [55.65870468861157]
This paper formulates unsupervised person ReID as a multi-label classification task to progressively seek true labels.
Our method starts by assigning each person image with a single-class label, then evolves to multi-label classification by leveraging the updated ReID model for label prediction.
To boost the ReID model training efficiency in multi-label classification, we propose the memory-based multi-label classification loss (MMCL)
arXiv Detail & Related papers (2020-04-20T12:13:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.