All Labels Together: Low-shot Intent Detection with an Efficient Label
Semantic Encoding Paradigm
- URL: http://arxiv.org/abs/2309.03563v2
- Date: Fri, 8 Sep 2023 01:00:52 GMT
- Title: All Labels Together: Low-shot Intent Detection with an Efficient Label
Semantic Encoding Paradigm
- Authors: Jiangshu Du, Congying Xia, Wenpeng Yin, Tingting Liang, Philip S. Yu
- Abstract summary: In intent detection tasks, leveraging meaningful semantic information from intent labels can be particularly beneficial for few-shot scenarios.
We present an end-to-end One-to-All system that enables the comparison of an input utterance with all label candidates.
Experiments on three few-shot intent detection tasks demonstrate that One-to-All is especially effective when the training resource is extremely scarce.
- Score: 48.02790193676742
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In intent detection tasks, leveraging meaningful semantic information from
intent labels can be particularly beneficial for few-shot scenarios. However,
existing few-shot intent detection methods either ignore the intent labels,
(e.g. treating intents as indices) or do not fully utilize this information
(e.g. only using part of the intent labels). In this work, we present an
end-to-end One-to-All system that enables the comparison of an input utterance
with all label candidates. The system can then fully utilize label semantics in
this way. Experiments on three few-shot intent detection tasks demonstrate that
One-to-All is especially effective when the training resource is extremely
scarce, achieving state-of-the-art performance in 1-, 3- and 5-shot settings.
Moreover, we present a novel pretraining strategy for our model that utilizes
indirect supervision from paraphrasing, enabling zero-shot cross-domain
generalization on intent detection tasks. Our code is at
https://github.com/jiangshdd/AllLablesTogether.
Related papers
- IDAS: Intent Discovery with Abstractive Summarization [16.731183915325584]
We show that recent competitive methods in intent discovery can be outperformed by clustering utterances based on abstractive summaries.
We contribute the IDAS approach, which collects a set of descriptive utterance labels by prompting a Large Language Model.
The utterances and their resulting noisy labels are then encoded by a frozen pre-trained encoder, and subsequently clustered to recover the latent intents.
arXiv Detail & Related papers (2023-05-31T12:19:40Z) - New Intent Discovery with Pre-training and Contrastive Learning [21.25371293641141]
New intent discovery aims to uncover novel intent categories from user utterances to expand the set of supported intent classes.
Existing approaches typically rely on a large amount of labeled utterances.
We propose a new contrastive loss to exploit self-supervisory signals in unlabeled data for clustering.
arXiv Detail & Related papers (2022-05-25T17:07:25Z) - Learning to Detect Mobile Objects from LiDAR Scans Without Labels [60.49869345286879]
Current 3D object detectors for autonomous driving are almost entirely trained on human-annotated data.
This paper proposes an alternative approach based on unlabeled data, which can be collected cheaply and in abundance almost everywhere on earth.
arXiv Detail & Related papers (2022-03-29T20:05:24Z) - Exploring the Limits of Natural Language Inference Based Setup for
Few-Shot Intent Detection [13.971616443394474]
Generalized Few-shot intent detection is more realistic but challenging setup.
We employ a simple and effective method based on Natural Language Inference.
Our method achieves state-of-the-art results on 1-shot and 5-shot intent detection task.
arXiv Detail & Related papers (2021-12-14T14:47:23Z) - Few-Shot Intent Detection via Contrastive Pre-Training and Fine-Tuning [27.154414939086426]
We present a simple yet effective few-shot intent detection schema via contrastive pre-training and fine-tuning.
We first conduct self-supervised contrastive pre-training on collected intent datasets, which implicitly learns to discriminate semantically similar utterances.
We then perform few-shot intent detection together with supervised contrastive learning, which explicitly pulls utterances from the same intent closer.
arXiv Detail & Related papers (2021-09-13T22:28:58Z) - Discovering New Intents with Deep Aligned Clustering [19.11073686645496]
We propose an effective method, Deep Aligned Clustering, to discover new intents with the aid of limited known intent data.
With an unknown number of new intents, we predict the number of intent categories by eliminating low-confidence intent-wise clusters.
Experiments on two benchmark datasets show that our method is more robust and achieves substantial improvements over the state-of-the-art methods.
arXiv Detail & Related papers (2020-12-16T14:32:06Z) - Few-shot Learning for Multi-label Intent Detection [59.66787898744991]
State-of-the-art work estimates label-instance relevance scores and uses a threshold to select multiple associated intent labels.
Experiments on two datasets show that the proposed model significantly outperforms strong baselines in both one-shot and five-shot settings.
arXiv Detail & Related papers (2020-10-11T14:42:18Z) - Learning with Weak Supervision for Email Intent Detection [56.71599262462638]
We propose to leverage user actions as a source of weak supervision to detect intents in emails.
We develop an end-to-end robust deep neural network model for email intent identification.
arXiv Detail & Related papers (2020-05-26T23:41:05Z) - AGIF: An Adaptive Graph-Interactive Framework for Joint Multiple Intent
Detection and Slot Filling [69.59096090788125]
In this paper, we propose an Adaptive Graph-Interactive Framework (AGIF) for joint multiple intent detection and slot filling.
We introduce an intent-slot graph interaction layer to model the strong correlation between the slot and intents.
Such an interaction layer is applied to each token adaptively, which has the advantage to automatically extract the relevant intents information.
arXiv Detail & Related papers (2020-04-21T15:07:34Z) - Efficient Intent Detection with Dual Sentence Encoders [53.16532285820849]
We introduce intent detection methods backed by pretrained dual sentence encoders such as USE and ConveRT.
We demonstrate the usefulness and wide applicability of the proposed intent detectors, showing that they outperform intent detectors based on fine-tuning the full BERT-Large model.
We release our code, as well as a new challenging single-domain intent detection dataset.
arXiv Detail & Related papers (2020-03-10T15:33:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.