Automatic Intent-Slot Induction for Dialogue Systems
- URL: http://arxiv.org/abs/2103.08886v1
- Date: Tue, 16 Mar 2021 07:21:31 GMT
- Title: Automatic Intent-Slot Induction for Dialogue Systems
- Authors: Zengfeng Zeng, Dan Ma, Haiqin Yang, Zhen Gou and Jianping Shen
- Abstract summary: We propose a new task of em automatic intent-slot induction and propose a novel domain-independent tool.
That is, we design a coarse-to-fine three-step procedure including role-labeling, Concept-mining, And Pattern-mining.
We show that our RCAP can generate satisfactory SLU schema and outperforms the state-of-the-art supervised learning method.
- Score: 5.6195418981579435
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Automatically and accurately identifying user intents and filling the
associated slots from their spoken language are critical to the success of
dialogue systems. Traditional methods require manually defining the
DOMAIN-INTENT-SLOT schema and asking many domain experts to annotate the
corresponding utterances, upon which neural models are trained. This procedure
brings the challenges of information sharing hindering, out-of-schema, or data
sparsity in open-domain dialogue systems. To tackle these challenges, we
explore a new task of {\em automatic intent-slot induction} and propose a novel
domain-independent tool. That is, we design a coarse-to-fine three-step
procedure including Role-labeling, Concept-mining, And Pattern-mining (RCAP):
(1) role-labeling: extracting keyphrases from users' utterances and classifying
them into a quadruple of coarsely-defined intent-roles via sequence labeling;
(2) concept-mining: clustering the extracted intent-role mentions and naming
them into abstract fine-grained concepts; (3) pattern-mining: applying the
Apriori algorithm to mine intent-role patterns and automatically inferring the
intent-slot using these coarse-grained intent-role labels and fine-grained
concepts. Empirical evaluations on both real-world in-domain and out-of-domain
datasets show that: (1) our RCAP can generate satisfactory SLU schema and
outperforms the state-of-the-art supervised learning method; (2) our RCAP can
be directly applied to out-of-domain datasets and gain at least 76\%
improvement of F1-score on intent detection and 41\% improvement of F1-score on
slot filling; (3) our RCAP exhibits its power in generic intent-slot
extractions with less manual effort, which opens pathways for schema induction
on new domains and unseen intent-slot discovery for generalizable dialogue
systems.
Related papers
- EAGER: Two-Stream Generative Recommender with Behavior-Semantic Collaboration [63.112790050749695]
We introduce EAGER, a novel generative recommendation framework that seamlessly integrates both behavioral and semantic information.
We validate the effectiveness of EAGER on four public benchmarks, demonstrating its superior performance compared to existing methods.
arXiv Detail & Related papers (2024-06-20T06:21:56Z) - A BiRGAT Model for Multi-intent Spoken Language Understanding with
Hierarchical Semantic Frames [30.200413352223347]
We first propose a Multi-Intent dataset which is collected from a realistic in-Vehicle dialogue System, called MIVS.
The target semantic frame is organized in a 3-layer hierarchical structure to tackle the alignment and assignment problems in multi-intent cases.
We devise a BiRGAT model to encode the hierarchy of items, the backbone of which is a dual relational graph attention network.
arXiv Detail & Related papers (2024-02-28T11:39:26Z) - Slot Induction via Pre-trained Language Model Probing and Multi-level
Contrastive Learning [62.839109775887025]
Slot Induction (SI) task whose objective is to induce slot boundaries without explicit knowledge of token-level slot annotations.
We propose leveraging Unsupervised Pre-trained Language Model (PLM) Probing and Contrastive Learning mechanism to exploit unsupervised semantic knowledge extracted from PLM.
Our approach is shown to be effective in SI task and capable of bridging the gaps with token-level supervised models on two NLU benchmark datasets.
arXiv Detail & Related papers (2023-08-09T05:08:57Z) - Zero-Shot Generalizable End-to-End Task-Oriented Dialog System using
Context Summarization and Domain Schema [2.7178968279054936]
State-of-the-art approaches in task-oriented dialog systems formulate the problem as a conditional sequence generation task.
This requires labeled training data for each new domain or task.
We introduce a novel Zero-Shot generalizable end-to-end Task-oriented Dialog system, ZS-ToD.
arXiv Detail & Related papers (2023-03-28T18:56:31Z) - DORIC : Domain Robust Fine-Tuning for Open Intent Clustering through
Dependency Parsing [14.709084509818474]
DSTC11-Track2 aims to provide a benchmark for zero-shot, cross-domain, intent-set induction.
We leveraged a multi-domain dialogue dataset to fine-tune the language model and proposed extracting Verb-Object pairs.
Our approach achieved 3rd place in the precision score and showed superior accuracy and normalized mutual information (NMI) score than the baseline model.
arXiv Detail & Related papers (2023-03-17T08:12:36Z) - Self-Paced Learning for Open-Set Domain Adaptation [50.620824701934]
Traditional domain adaptation methods presume that the classes in the source and target domains are identical.
Open-set domain adaptation (OSDA) addresses this limitation by allowing previously unseen classes in the target domain.
We propose a novel framework based on self-paced learning to distinguish common and unknown class samples.
arXiv Detail & Related papers (2023-03-10T14:11:09Z) - Hierarchical Modeling for Out-of-Scope Domain and Intent Classification [55.23920796595698]
This paper focuses on out-of-scope intent classification in dialog systems.
We propose a hierarchical multi-task learning approach based on a joint model to classify domain and intent simultaneously.
Experiments show that the model outperforms existing methods in terms of accuracy, out-of-scope recall and F1.
arXiv Detail & Related papers (2021-04-30T06:38:23Z) - PIN: A Novel Parallel Interactive Network for Spoken Language
Understanding [68.53121591998483]
In the existing RNN-based approaches, ID and SF tasks are often jointly modeled to utilize the correlation information between them.
The experiments on two benchmark datasets, i.e., SNIPS and ATIS, demonstrate the effectiveness of our approach.
More encouragingly, by using the feature embedding of the utterance generated by the pre-trained language model BERT, our method achieves the state-of-the-art among all comparison approaches.
arXiv Detail & Related papers (2020-09-28T15:59:31Z) - AGIF: An Adaptive Graph-Interactive Framework for Joint Multiple Intent
Detection and Slot Filling [69.59096090788125]
In this paper, we propose an Adaptive Graph-Interactive Framework (AGIF) for joint multiple intent detection and slot filling.
We introduce an intent-slot graph interaction layer to model the strong correlation between the slot and intents.
Such an interaction layer is applied to each token adaptively, which has the advantage to automatically extract the relevant intents information.
arXiv Detail & Related papers (2020-04-21T15:07:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.