Learning Disentangled Intent Representations for Zero-shot Intent
Detection
- URL: http://arxiv.org/abs/2012.01721v1
- Date: Thu, 3 Dec 2020 06:41:09 GMT
- Title: Learning Disentangled Intent Representations for Zero-shot Intent
Detection
- Authors: Qingyi Si, Yuanxin Liu, Peng Fu, Jiangnan Li, Zheng Lin and Weiping
Wang
- Abstract summary: We propose a class-transductive framework that utilizes unseen class labels to learn Disentangled Representations (DIR)
Under this framework, we introduce a multi-task learning objective, which encourages the model to learn the distinctions among intents.
Experiments on two real-world datasets show that the proposed framework brings consistent improvement to the baseline systems.
- Score: 13.19024497857648
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Zero-shot intent detection (ZSID) aims to deal with the continuously emerging
intents without annotated training data. However, existing ZSID systems suffer
from two limitations: 1) They are not good at modeling the relationship between
seen and unseen intents, when the label names are given in the form of raw
phrases or sentences. 2) They cannot effectively recognize unseen intents under
the generalized intent detection (GZSID) setting. A critical factor behind
these limitations is the representations of unseen intents, which cannot be
learned in the training stage. To address this problem, we propose a
class-transductive framework that utilizes unseen class labels to learn
Disentangled Intent Representations (DIR). Specifically, we allow the model to
predict unseen intents in the training stage, with the corresponding label
names serving as input utterances. Under this framework, we introduce a
multi-task learning objective, which encourages the model to learn the
distinctions among intents, and a similarity scorer, which estimates the
connections among intents more accurately based on the learned intent
representations. Since the purpose of DIR is to provide better intent
representations, it can be easily integrated with existing ZSID and GZSID
methods. Experiments on two real-world datasets show that the proposed
framework brings consistent improvement to the baseline systems, regardless of
the model architectures or zero-shot learning strategies.
Related papers
- Pseudo-Label Enhanced Prototypical Contrastive Learning for Uniformed Intent Discovery [27.18799732585361]
We propose a Pseudo-Label enhanced Prototypical Contrastive Learning (PLPCL) model for uniformed intent discovery.
We iteratively utilize pseudo-labels to explore potential positive/negative samples for contrastive learning and bridge the gap between representation and clustering.
Our method has been proven effective in two different settings of discovering new intents.
arXiv Detail & Related papers (2024-10-26T16:22:45Z) - Label Aware Speech Representation Learning For Language Identification [49.197215416945596]
We propose a novel framework of combining self-supervised representation learning with the language label information for the pre-training task.
This framework, termed as Label Aware Speech Representation (LASR) learning, uses a triplet based objective function to incorporate language labels along with the self-supervised loss function.
arXiv Detail & Related papers (2023-06-07T12:14:16Z) - Zero-Shot Learning for Joint Intent and Slot Labeling [11.82805641934772]
We show that one can profitably perform joint zero-shot intent classification and slot labeling.
We describe NN architectures that translate between word and sentence embedding spaces.
arXiv Detail & Related papers (2022-11-29T01:58:25Z) - New Intent Discovery with Pre-training and Contrastive Learning [21.25371293641141]
New intent discovery aims to uncover novel intent categories from user utterances to expand the set of supported intent classes.
Existing approaches typically rely on a large amount of labeled utterances.
We propose a new contrastive loss to exploit self-supervisory signals in unlabeled data for clustering.
arXiv Detail & Related papers (2022-05-25T17:07:25Z) - Cross-modal Representation Learning for Zero-shot Action Recognition [67.57406812235767]
We present a cross-modal Transformer-based framework, which jointly encodes video data and text labels for zero-shot action recognition (ZSAR)
Our model employs a conceptually new pipeline by which visual representations are learned in conjunction with visual-semantic associations in an end-to-end manner.
Experiment results show our model considerably improves upon the state of the arts in ZSAR, reaching encouraging top-1 accuracy on UCF101, HMDB51, and ActivityNet benchmark datasets.
arXiv Detail & Related papers (2022-05-03T17:39:27Z) - Intent Contrastive Learning for Sequential Recommendation [86.54439927038968]
We introduce a latent variable to represent users' intents and learn the distribution function of the latent variable via clustering.
We propose to leverage the learned intents into SR models via contrastive SSL, which maximizes the agreement between a view of sequence and its corresponding intent.
Experiments conducted on four real-world datasets demonstrate the superiority of the proposed learning paradigm.
arXiv Detail & Related papers (2022-02-05T09:24:13Z) - An Explicit-Joint and Supervised-Contrastive Learning Framework for
Few-Shot Intent Classification and Slot Filling [12.85364483952161]
Intent classification (IC) and slot filling (SF) are critical building blocks in task-oriented dialogue systems.
Few IC/SF models perform well when the number of training samples per class is quite small.
We propose a novel explicit-joint and supervised-contrastive learning framework for few-shot intent classification and slot filling.
arXiv Detail & Related papers (2021-10-26T13:28:28Z) - Semantics-Guided Contrastive Network for Zero-Shot Object detection [67.61512036994458]
Zero-shot object detection (ZSD) is a new challenge in computer vision.
We develop ContrastZSD, a framework that brings contrastive learning mechanism into the realm of zero-shot detection.
Our method outperforms the previous state-of-the-art on both ZSD and generalized ZSD tasks.
arXiv Detail & Related papers (2021-09-04T03:32:15Z) - Generalized Zero-shot Intent Detection via Commonsense Knowledge [5.398580049917152]
We propose RIDE: an intent detection model that leverages commonsense knowledge in an unsupervised fashion to overcome the issue of training data scarcity.
RIDE computes robust and generalizable relationship meta-features that capture deep semantic relationships between utterances and intent labels.
Our extensive experimental analysis on three widely-used intent detection benchmarks shows that relationship meta-features significantly increase the accuracy of detecting both seen and unseen intents.
arXiv Detail & Related papers (2021-02-04T23:36:41Z) - Discriminative Nearest Neighbor Few-Shot Intent Detection by
Transferring Natural Language Inference [150.07326223077405]
Few-shot learning is attracting much attention to mitigate data scarcity.
We present a discriminative nearest neighbor classification with deep self-attention.
We propose to boost the discriminative ability by transferring a natural language inference (NLI) model.
arXiv Detail & Related papers (2020-10-25T00:39:32Z) - Efficient Intent Detection with Dual Sentence Encoders [53.16532285820849]
We introduce intent detection methods backed by pretrained dual sentence encoders such as USE and ConveRT.
We demonstrate the usefulness and wide applicability of the proposed intent detectors, showing that they outperform intent detectors based on fine-tuning the full BERT-Large model.
We release our code, as well as a new challenging single-domain intent detection dataset.
arXiv Detail & Related papers (2020-03-10T15:33:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.