Multi-grained Label Refinement Network with Dependency Structures for
Joint Intent Detection and Slot Filling
- URL: http://arxiv.org/abs/2209.04156v1
- Date: Fri, 9 Sep 2022 07:27:38 GMT
- Title: Multi-grained Label Refinement Network with Dependency Structures for
Joint Intent Detection and Slot Filling
- Authors: Baohang Zhou, Ying Zhang, Xuhui Sui, Kehui Song, Xiaojie Yuan
- Abstract summary: intent and semantic components of a utterance are dependent on the syntactic elements of a sentence.
In this paper, we investigate a multi-grained label refinement network, which utilizes dependency structures and label semantic embeddings.
Considering to enhance syntactic representations, we introduce the dependency structures of sentences into our model by graph attention layer.
- Score: 13.963083174197164
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Slot filling and intent detection are two fundamental tasks in the field of
natural language understanding. Due to the strong correlation between these two
tasks, previous studies make efforts on modeling them with multi-task learning
or designing feature interaction modules to improve the performance of each
task. However, none of the existing approaches consider the relevance between
the structural information of sentences and the label semantics of two tasks.
The intent and semantic components of a utterance are dependent on the
syntactic elements of a sentence. In this paper, we investigate a multi-grained
label refinement network, which utilizes dependency structures and label
semantic embeddings. Considering to enhance syntactic representations, we
introduce the dependency structures of sentences into our model by graph
attention layer. To capture the semantic dependency between the syntactic
information and task labels, we combine the task specific features with
corresponding label embeddings by attention mechanism. The experimental results
demonstrate that our model achieves the competitive performance on two public
datasets.
Related papers
- Entity-Aware Self-Attention and Contextualized GCN for Enhanced Relation Extraction in Long Sentences [5.453850739960517]
We propose a novel model, Entity-aware Self-attention Contextualized GCN (ESC-GCN), which efficiently incorporates syntactic structure of input sentences and semantic context of sequences.
Our model achieves encouraging performance as compared to existing dependency-based and sequence-based models.
arXiv Detail & Related papers (2024-09-15T10:50:51Z) - A Unified Framework for Multi-intent Spoken Language Understanding with
prompting [14.17726194025463]
We describe a Prompt-based Spoken Language Understanding (PromptSLU) framework, to intuitively unify two sub-tasks into the same form.
In detail, ID and SF are completed by concisely filling the utterance into task-specific prompt templates as input, and sharing output formats of key-value pairs sequence.
Experiment results show that our framework outperforms several state-of-the-art baselines on two public datasets.
arXiv Detail & Related papers (2022-10-07T05:58:05Z) - Coarse-to-Fine: Hierarchical Multi-task Learning for Natural Language
Understanding [51.31622274823167]
We propose a hierarchical framework with a coarse-to-fine paradigm, with the bottom level shared to all the tasks, the mid-level divided to different groups, and the top-level assigned to each of the tasks.
This allows our model to learn basic language properties from all tasks, boost performance on relevant tasks, and reduce the negative impact from irrelevant tasks.
arXiv Detail & Related papers (2022-08-19T02:46:20Z) - Compositional Generalization in Grounded Language Learning via Induced
Model Sparsity [81.38804205212425]
We consider simple language-conditioned navigation problems in a grid world environment with disentangled observations.
We design an agent that encourages sparse correlations between words in the instruction and attributes of objects, composing them together to find the goal.
Our agent maintains a high level of performance on goals containing novel combinations of properties even when learning from a handful of demonstrations.
arXiv Detail & Related papers (2022-07-06T08:46:27Z) - Unified Semantic Typing with Meaningful Label Inference [19.308286513235934]
We present UniST, a unified framework for semantic typing.
UniST captures label semantics by projecting both inputs and labels into a joint semantic embedding space.
Our experiments demonstrate that UniST achieves strong performance across three semantic typing tasks.
arXiv Detail & Related papers (2022-05-04T00:28:17Z) - Transformer-based Dual Relation Graph for Multi-label Image Recognition [56.12543717723385]
We propose a novel Transformer-based Dual Relation learning framework.
We explore two aspects of correlation, i.e., structural relation graph and semantic relation graph.
Our approach achieves new state-of-the-art on two popular multi-label recognition benchmarks.
arXiv Detail & Related papers (2021-10-10T07:14:52Z) - Context-Aware Interaction Network for Question Matching [51.76812857301819]
We propose a context-aware interaction network (COIN) to align two sequences and infer their semantic relationship.
Specifically, each interaction block includes (1) a context-aware cross-attention mechanism to effectively integrate contextual information, and (2) a gate fusion layer to flexibly interpolate aligned representations.
arXiv Detail & Related papers (2021-04-17T05:03:56Z) - Multilingual Irony Detection with Dependency Syntax and Neural Models [61.32653485523036]
It focuses on the contribution from syntactic knowledge, exploiting linguistic resources where syntax is annotated according to the Universal Dependencies scheme.
The results suggest that fine-grained dependency-based syntactic information is informative for the detection of irony.
arXiv Detail & Related papers (2020-11-11T11:22:05Z) - Joint Semantic Analysis with Document-Level Cross-Task Coherence Rewards [13.753240692520098]
We present a neural network architecture for joint coreference resolution and semantic role labeling for English.
We use reinforcement learning to encourage global coherence over the document and between semantic annotations.
This leads to improvements on both tasks in multiple datasets from different domains.
arXiv Detail & Related papers (2020-10-12T09:36:24Z) - Bidirectional Graph Reasoning Network for Panoptic Segmentation [126.06251745669107]
We introduce a Bidirectional Graph Reasoning Network (BGRNet) to mine the intra-modular and intermodular relations within and between foreground things and background stuff classes.
BGRNet first constructs image-specific graphs in both instance and semantic segmentation branches that enable flexible reasoning at the proposal level and class level.
arXiv Detail & Related papers (2020-04-14T02:32:10Z) - Edge-Enhanced Graph Convolution Networks for Event Detection with
Syntactic Relation [14.823029993354794]
Event detection is a key subtask of information extraction.
We propose a novel architecture named Edge-Enhanced Graph Convolution Networks (EE-GCN)
EE-GCN simultaneously exploits syntactic structure and typed dependency label information to perform ED.
arXiv Detail & Related papers (2020-02-25T09:18:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.