Continuous representations of intents for dialogue systems
- URL: http://arxiv.org/abs/2105.03716v1
- Date: Sat, 8 May 2021 15:08:20 GMT
- Title: Continuous representations of intents for dialogue systems
- Authors: Sindre Andr\'e Jacobsen and Anton Ragni
- Abstract summary: Up until recently the focus has been on detecting a fixed, discrete, number of seen intents.
Recent years have seen some work done on unseen intent detection in the context of zero-shot learning.
This paper proposes a novel model where intents are continuous points placed in a specialist Intent Space.
- Score: 10.031004070657122
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Intent modelling has become an important part of modern dialogue systems.
With the rapid expansion of practical dialogue systems and virtual assistants,
such as Amazon Alexa, Apple Siri, and Google Assistant, the interest has only
increased. However, up until recently the focus has been on detecting a fixed,
discrete, number of seen intents. Recent years have seen some work done on
unseen intent detection in the context of zero-shot learning. This paper
continues the prior work by proposing a novel model where intents are
continuous points placed in a specialist Intent Space that yields several
advantages. First, the continuous representation enables to investigate
relationships between the seen intents. Second, it allows any unseen intent to
be reliably represented given limited quantities of data. Finally, this paper
will show how the proposed model can be augmented with unseen intents without
retraining any of the seen ones. Experiments show that the model can reliably
add unseen intents with a high accuracy while retaining a high performance on
the seen intents.
Related papers
- Discovering New Intents Using Latent Variables [51.50374666602328]
We propose a probabilistic framework for discovering intents where intent assignments are treated as latent variables.
In E-step, we conduct discovering intents and explore the intrinsic structure of unlabeled data by the posterior of intent assignments.
In M-step, we alleviate the forgetting of prior knowledge transferred from known intents by optimizing the discrimination of labeled data.
arXiv Detail & Related papers (2022-10-21T08:29:45Z) - Zero-Shot Prompting for Implicit Intent Prediction and Recommendation
with Commonsense Reasoning [28.441725610692714]
This paper proposes a framework of multi-domain dialogue systems, which can automatically infer implicit intents based on user utterances.
The proposed framework is demonstrated effective to realize implicit intents and recommend associated bots in a zero-shot manner.
arXiv Detail & Related papers (2022-10-12T03:33:49Z) - Template-based Approach to Zero-shot Intent Recognition [7.330908962006392]
In this paper, we explore the generalized zero-shot setup for intent recognition.
Following best practices for zero-shot text classification, we treat the task with a sentence pair modeling approach.
We outperform previous state-of-the-art f1-measure by up to 16% for unseen intents.
arXiv Detail & Related papers (2022-06-22T08:44:59Z) - A Framework to Generate High-Quality Datapoints for Multiple Novel
Intent Detection [24.14668837496296]
MNID is a framework to detect multiple novel intents with budgeted human annotation cost.
It outperforms the baseline methods in terms of accuracy and F1-score.
arXiv Detail & Related papers (2022-05-04T11:32:15Z) - Generalized Zero-shot Intent Detection via Commonsense Knowledge [5.398580049917152]
We propose RIDE: an intent detection model that leverages commonsense knowledge in an unsupervised fashion to overcome the issue of training data scarcity.
RIDE computes robust and generalizable relationship meta-features that capture deep semantic relationships between utterances and intent labels.
Our extensive experimental analysis on three widely-used intent detection benchmarks shows that relationship meta-features significantly increase the accuracy of detecting both seen and unseen intents.
arXiv Detail & Related papers (2021-02-04T23:36:41Z) - Discriminative Nearest Neighbor Few-Shot Intent Detection by
Transferring Natural Language Inference [150.07326223077405]
Few-shot learning is attracting much attention to mitigate data scarcity.
We present a discriminative nearest neighbor classification with deep self-attention.
We propose to boost the discriminative ability by transferring a natural language inference (NLI) model.
arXiv Detail & Related papers (2020-10-25T00:39:32Z) - Learning Long-term Visual Dynamics with Region Proposal Interaction
Networks [75.06423516419862]
We build object representations that can capture inter-object and object-environment interactions over a long-range.
Thanks to the simple yet effective object representation, our approach outperforms prior methods by a significant margin.
arXiv Detail & Related papers (2020-08-05T17:48:00Z) - Learning with Weak Supervision for Email Intent Detection [56.71599262462638]
We propose to leverage user actions as a source of weak supervision to detect intents in emails.
We develop an end-to-end robust deep neural network model for email intent identification.
arXiv Detail & Related papers (2020-05-26T23:41:05Z) - AGIF: An Adaptive Graph-Interactive Framework for Joint Multiple Intent
Detection and Slot Filling [69.59096090788125]
In this paper, we propose an Adaptive Graph-Interactive Framework (AGIF) for joint multiple intent detection and slot filling.
We introduce an intent-slot graph interaction layer to model the strong correlation between the slot and intents.
Such an interaction layer is applied to each token adaptively, which has the advantage to automatically extract the relevant intents information.
arXiv Detail & Related papers (2020-04-21T15:07:34Z) - Efficient Intent Detection with Dual Sentence Encoders [53.16532285820849]
We introduce intent detection methods backed by pretrained dual sentence encoders such as USE and ConveRT.
We demonstrate the usefulness and wide applicability of the proposed intent detectors, showing that they outperform intent detectors based on fine-tuning the full BERT-Large model.
We release our code, as well as a new challenging single-domain intent detection dataset.
arXiv Detail & Related papers (2020-03-10T15:33:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.