JPIS: A Joint Model for Profile-based Intent Detection and Slot Filling
with Slot-to-Intent Attention
- URL: http://arxiv.org/abs/2312.08737v2
- Date: Sat, 16 Dec 2023 14:50:53 GMT
- Title: JPIS: A Joint Model for Profile-based Intent Detection and Slot Filling
with Slot-to-Intent Attention
- Authors: Thinh Pham, Dat Quoc Nguyen
- Abstract summary: Profile-based intent detection and slot filling are important tasks aimed at reducing the ambiguity in user utterances.
We propose a joint model, namely JPIS, designed to enhance profile-based intent detection and slot filling.
Experimental results show that our JPIS substantially outperforms previous profile-based models.
- Score: 11.12158809959412
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Profile-based intent detection and slot filling are important tasks aimed at
reducing the ambiguity in user utterances by leveraging user-specific
supporting profile information. However, research in these two tasks has not
been extensively explored. To fill this gap, we propose a joint model, namely
JPIS, designed to enhance profile-based intent detection and slot filling. JPIS
incorporates the supporting profile information into its encoder and introduces
a slot-to-intent attention mechanism to transfer slot information
representations to intent detection. Experimental results show that our JPIS
substantially outperforms previous profile-based models, establishing a new
state-of-the-art performance in overall accuracy on the Chinese benchmark
dataset ProSLU.
Related papers
- MISCA: A Joint Model for Multiple Intent Detection and Slot Filling with
Intent-Slot Co-Attention [9.414164374919029]
Recent advanced approaches, which are joint models based on graphs, might still face two potential issues.
We propose a joint model named MISCA.
Our MISCA introduces an intent-slot co-attention mechanism and an underlying layer of label attention mechanism.
arXiv Detail & Related papers (2023-12-10T03:38:41Z) - Introducing "Forecast Utterance" for Conversational Data Science [2.3894779000840503]
This paper introduces a new concept called Forecast Utterance.
We then focus on the automatic and accurate interpretation of users' prediction goals from these utterances.
Specifically, we frame the task as a slot-filling problem, where each slot corresponds to a specific aspect of the goal prediction task.
We then employ two zero-shot methods for solving the slot-filling task, namely: 1) Entity Extraction (EE), and 2) Question-Answering (QA) techniques.
arXiv Detail & Related papers (2023-09-07T17:41:41Z) - Slot Induction via Pre-trained Language Model Probing and Multi-level
Contrastive Learning [62.839109775887025]
Slot Induction (SI) task whose objective is to induce slot boundaries without explicit knowledge of token-level slot annotations.
We propose leveraging Unsupervised Pre-trained Language Model (PLM) Probing and Contrastive Learning mechanism to exploit unsupervised semantic knowledge extracted from PLM.
Our approach is shown to be effective in SI task and capable of bridging the gaps with token-level supervised models on two NLU benchmark datasets.
arXiv Detail & Related papers (2023-08-09T05:08:57Z) - Robust Saliency-Aware Distillation for Few-shot Fine-grained Visual
Recognition [57.08108545219043]
Recognizing novel sub-categories with scarce samples is an essential and challenging research topic in computer vision.
Existing literature addresses this challenge by employing local-based representation approaches.
This article proposes a novel model, Robust Saliency-aware Distillation (RSaD), for few-shot fine-grained visual recognition.
arXiv Detail & Related papers (2023-05-12T00:13:17Z) - Explainable Slot Type Attentions to Improve Joint Intent Detection and
Slot Filling [39.22929726787844]
Joint intent detection and slot filling is a key research topic in natural language understanding (NLU)
Existing joint intent and slot filling systems analyze and compute features collectively for all slot types.
We propose a novel approach that: (i) learns to generate additional slot type specific features in order to improve accuracy and (ii) provides explanations for slot filling decisions for the first time in a joint NLU model.
arXiv Detail & Related papers (2022-10-19T00:56:10Z) - A Fast Attention Network for Joint Intent Detection and Slot Filling on
Edge Devices [5.982036462074887]
We propose a Fast Attention Network (FAN) for joint intent detection and slot filling tasks, guaranteeing both accuracy and latency.
FAN can be implemented on different encoders and delivers more accurate models at every speed level.
Our experiments on the Jetson Nano platform show that FAN inferences fifteen utterances per second with a small accuracy drop, showing its effectiveness and efficiency on edge devices.
arXiv Detail & Related papers (2022-05-16T13:06:51Z) - Information-Theoretic Odometry Learning [83.36195426897768]
We propose a unified information theoretic framework for learning-motivated methods aimed at odometry estimation.
The proposed framework provides an elegant tool for performance evaluation and understanding in information-theoretic language.
arXiv Detail & Related papers (2022-03-11T02:37:35Z) - Bi-directional Joint Neural Networks for Intent Classification and Slot
Filling [5.3361357265365035]
We propose a bi-directional joint model for intent classification and slot filling.
Our model achieves state-of-the-art results on intent classification accuracy, slot filling F1, and significantly improves sentence-level semantic frame accuracy.
arXiv Detail & Related papers (2022-02-26T06:35:21Z) - A Bayesian Framework for Information-Theoretic Probing [51.98576673620385]
We argue that probing should be seen as approximating a mutual information.
This led to the rather unintuitive conclusion that representations encode exactly the same information about a target task as the original sentences.
This paper proposes a new framework to measure what we term Bayesian mutual information.
arXiv Detail & Related papers (2021-09-08T18:08:36Z) - PIN: A Novel Parallel Interactive Network for Spoken Language
Understanding [68.53121591998483]
In the existing RNN-based approaches, ID and SF tasks are often jointly modeled to utilize the correlation information between them.
The experiments on two benchmark datasets, i.e., SNIPS and ATIS, demonstrate the effectiveness of our approach.
More encouragingly, by using the feature embedding of the utterance generated by the pre-trained language model BERT, our method achieves the state-of-the-art among all comparison approaches.
arXiv Detail & Related papers (2020-09-28T15:59:31Z) - AGIF: An Adaptive Graph-Interactive Framework for Joint Multiple Intent
Detection and Slot Filling [69.59096090788125]
In this paper, we propose an Adaptive Graph-Interactive Framework (AGIF) for joint multiple intent detection and slot filling.
We introduce an intent-slot graph interaction layer to model the strong correlation between the slot and intents.
Such an interaction layer is applied to each token adaptively, which has the advantage to automatically extract the relevant intents information.
arXiv Detail & Related papers (2020-04-21T15:07:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.