Task Conditioned BERT for Joint Intent Detection and Slot-filling
- URL: http://arxiv.org/abs/2308.06165v1
- Date: Fri, 11 Aug 2023 14:47:27 GMT
- Title: Task Conditioned BERT for Joint Intent Detection and Slot-filling
- Authors: Diogo Tavares and Pedro Azevedo and David Semedo and Ricardo Sousa and
Jo\~ao Magalh\~aes
- Abstract summary: We investigate the hypothesis that solving these challenges as one unified model will allow the transfer of parameter support data across the different tasks.
The proposed principled model is based on a Transformer encoder, trained on multiple tasks, and leveraged by a rich input that conditions the model on the target inferences.
- Score: 4.564638655448509
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Dialogue systems need to deal with the unpredictability of user intents to
track dialogue state and the heterogeneity of slots to understand user
preferences. In this paper we investigate the hypothesis that solving these
challenges as one unified model will allow the transfer of parameter support
data across the different tasks. The proposed principled model is based on a
Transformer encoder, trained on multiple tasks, and leveraged by a rich input
that conditions the model on the target inferences. Conditioning the
Transformer encoder on multiple target inferences over the same corpus, i.e.,
intent and multiple slot types, allows learning richer language interactions
than a single-task model would be able to. In fact, experimental results
demonstrate that conditioning the model on an increasing number of dialogue
inference tasks leads to improved results: on the MultiWOZ dataset, the joint
intent and slot detection can be improved by 3.2\% by conditioning on intent,
10.8\% by conditioning on slot and 14.4\% by conditioning on both intent and
slots. Moreover, on real conversations with Farfetch costumers, the proposed
conditioned BERT can achieve high joint-goal and intent detection performance
throughout a dialogue.
Related papers
- Joint Multiple Intent Detection and Slot Filling with Supervised
Contrastive Learning and Self-Distillation [4.123763595394021]
Multiple intent detection and slot filling are fundamental and crucial tasks in spoken language understanding.
Joint models that can detect intents and extract slots simultaneously are preferred.
We present a method for multiple intent detection and slot filling by addressing these challenges.
arXiv Detail & Related papers (2023-08-28T15:36:33Z) - Gated Mechanism Enhanced Multi-Task Learning for Dialog Routing [28.870359916550996]
Gated Mechanism enhanced Multi-task Model (G3M)
Proposal includes a novel dialog encoder and two tailored gated mechanism modules.
Based on two datasets collected from real world applications, extensive experimental results demonstrate the effectiveness of our method.
arXiv Detail & Related papers (2023-04-07T16:51:46Z) - Exploring the Trade-off between Plausibility, Change Intensity and
Adversarial Power in Counterfactual Explanations using Multi-objective
Optimization [73.89239820192894]
We argue that automated counterfactual generation should regard several aspects of the produced adversarial instances.
We present a novel framework for the generation of counterfactual examples.
arXiv Detail & Related papers (2022-05-20T15:02:53Z) - Paragraph-based Transformer Pre-training for Multi-Sentence Inference [99.59693674455582]
We show that popular pre-trained transformers perform poorly when used for fine-tuning on multi-candidate inference tasks.
We then propose a new pre-training objective that models the paragraph-level semantics across multiple input sentences.
arXiv Detail & Related papers (2022-05-02T21:41:14Z) - Reasoning in Dialog: Improving Response Generation by Context Reading
Comprehension [49.92173751203827]
In multi-turn dialog, utterances do not always take the full form of sentences.
We propose to improve the response generation performance by examining the model's ability to answer a reading comprehension question.
arXiv Detail & Related papers (2020-12-14T10:58:01Z) - Example-Driven Intent Prediction with Observers [15.615065041164629]
We focus on the intent classification problem which aims to identify user intents given utterances addressed to the dialog system.
We propose two approaches for improving the generalizability of utterance classification models: (1) observers and (2) example-driven training.
arXiv Detail & Related papers (2020-10-17T01:03:06Z) - A Co-Interactive Transformer for Joint Slot Filling and Intent Detection [61.109486326954205]
Intent detection and slot filling are two main tasks for building a spoken language understanding (SLU) system.
Previous studies either model the two tasks separately or only consider the single information flow from intent to slot.
We propose a Co-Interactive Transformer to consider the cross-impact between the two tasks simultaneously.
arXiv Detail & Related papers (2020-10-08T10:16:52Z) - DCR-Net: A Deep Co-Interactive Relation Network for Joint Dialog Act
Recognition and Sentiment Classification [77.59549450705384]
In dialog system, dialog act recognition and sentiment classification are two correlative tasks.
Most of the existing systems either treat them as separate tasks or just jointly model the two tasks.
We propose a Deep Co-Interactive Relation Network (DCR-Net) to explicitly consider the cross-impact and model the interaction between the two tasks.
arXiv Detail & Related papers (2020-08-16T14:13:32Z) - Efficient Intent Detection with Dual Sentence Encoders [53.16532285820849]
We introduce intent detection methods backed by pretrained dual sentence encoders such as USE and ConveRT.
We demonstrate the usefulness and wide applicability of the proposed intent detectors, showing that they outperform intent detectors based on fine-tuning the full BERT-Large model.
We release our code, as well as a new challenging single-domain intent detection dataset.
arXiv Detail & Related papers (2020-03-10T15:33:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.