Lexicon-injected Semantic Parsing for Task-Oriented Dialog
- URL: http://arxiv.org/abs/2211.14508v1
- Date: Sat, 26 Nov 2022 07:59:20 GMT
- Title: Lexicon-injected Semantic Parsing for Task-Oriented Dialog
- Authors: Xiaojun Meng, Wenlin Dai, Yasheng Wang, Baojun Wang, Zhiyong Wu, Xin
Jiang, Qun Liu
- Abstract summary: We present a novel lexicon- semanticinjected, which collects slot labels of tree representation as a lexicon and injects lexical features to the span representation of the tree nodes.
Our best result produces a new state-of-the-art result (87.62%) on the TOP dataset, and demonstrates its adaptability to frequently updated slot lexicon entries in real task-oriented dialog.
- Score: 31.42253032456493
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Recently, semantic parsing using hierarchical representations for dialog
systems has captured substantial attention. Task-Oriented Parse (TOP), a tree
representation with intents and slots as labels of nested tree nodes, has been
proposed for parsing user utterances. Previous TOP parsing methods are limited
on tackling unseen dynamic slot values (e.g., new songs and locations added),
which is an urgent matter for real dialog systems. To mitigate this issue, we
first propose a novel span-splitting representation for span-based parser that
outperforms existing methods. Then we present a novel lexicon-injected semantic
parser, which collects slot labels of tree representation as a lexicon, and
injects lexical features to the span representation of parser. An additional
slot disambiguation technique is involved to remove inappropriate span match
occurrences from the lexicon. Our best parser produces a new state-of-the-art
result (87.62%) on the TOP dataset, and demonstrates its adaptability to
frequently updated slot lexicon entries in real task-oriented dialog, with no
need of retraining.
Related papers
- Structured Dialogue Discourse Parsing [79.37200787463917]
discourse parsing aims to uncover the internal structure of a multi-participant conversation.
We propose a principled method that improves upon previous work from two perspectives: encoding and decoding.
Experiments show that our method achieves new state-of-the-art, surpassing the previous model by 2.3 on STAC and 1.5 on Molweni.
arXiv Detail & Related papers (2023-06-26T22:51:01Z) - Conversational Semantic Parsing using Dynamic Context Graphs [68.72121830563906]
We consider the task of conversational semantic parsing over general purpose knowledge graphs (KGs) with millions of entities, and thousands of relation-types.
We focus on models which are capable of interactively mapping user utterances into executable logical forms.
arXiv Detail & Related papers (2023-05-04T16:04:41Z) - PIZZA: A new benchmark for complex end-to-end task-oriented parsing [3.5106870325869886]
This paper introduces a new dataset for parsing pizza and drink orders, whose semantics cannot be captured by flat slots and intents.
We perform an evaluation of deep-learning techniques for task-oriented parsing on this dataset, including different flavors of seq2seqNGs.
arXiv Detail & Related papers (2022-12-01T04:20:07Z) - SPACE-2: Tree-Structured Semi-Supervised Contrastive Pre-training for
Task-Oriented Dialog Understanding [68.94808536012371]
We propose a tree-structured pre-trained conversation model, which learns dialog representations from limited labeled dialogs and large-scale unlabeled dialog corpora.
Our method can achieve new state-of-the-art results on the DialoGLUE benchmark consisting of seven datasets and four popular dialog understanding tasks.
arXiv Detail & Related papers (2022-09-14T13:42:50Z) - DP-Parse: Finding Word Boundaries from Raw Speech with an Instance
Lexicon [18.05179713472479]
We introduce DP-Parse, which uses similar principles but only relies on an instance lexicon of word tokens.
On the Zero Resource Speech Benchmark 2017, our model sets a new speech segmentation state-of-the-art in 5 languages.
Despite lacking a type lexicon, DP-Parse can be pipelined to a language model and learn and semantic representations as assessed by a new spoken word embedding benchmark.
arXiv Detail & Related papers (2022-06-22T19:15:57Z) - Hierarchical Context Tagging for Utterance Rewriting [51.251400047377324]
Methods that tag rather than linearly generate sequences have proven stronger in both in- and out-of-domain rewriting settings.
We propose a hierarchical context tagger that mitigates this issue by predicting slotted rules.
Experiments on several benchmarks show that HCT can outperform state-of-the-art rewriting systems by 2 BLEU points.
arXiv Detail & Related papers (2022-06-22T17:09:34Z) - LEXpander: applying colexification networks to automated lexicon
expansion [0.16804697591495946]
We present LEXpander, a method for lexicon expansion that leverages novel data on colexification.
We find that LEXpander outperforms existing approaches in terms of both precision and the trade-off between precision and recall of generated word lists.
arXiv Detail & Related papers (2022-05-31T14:55:29Z) - DyLex: Incorporating Dynamic Lexicons into BERT for Sequence Labeling [49.3379730319246]
We propose DyLex, a plug-in lexicon incorporation approach for BERT based sequence labeling tasks.
We adopt word-agnostic tag embeddings to avoid re-training the representation while updating the lexicon.
Finally, we introduce a col-wise attention based knowledge fusion mechanism to guarantee the pluggability of the proposed framework.
arXiv Detail & Related papers (2021-09-18T03:15:49Z) - Semantic Parsing in Task-Oriented Dialog with Recursive Insertion-based
Encoder [6.507504084891086]
We introduce a Recursive INsertion-based entity recognition (RINE) approach for semantic parsing in task-oriented dialog.
RINE achieves state-of-the-art exact match accuracy on low- and high-resource versions of the conversational semantic parsing benchmark TOP.
Our approach is 2-3.5 times faster than the sequence-to-sequence model at inference time.
arXiv Detail & Related papers (2021-09-09T18:23:45Z) - RST Parsing from Scratch [14.548146390081778]
We introduce a novel end-to-end formulation of document-level discourse parsing in the Rhetorical Structure Theory (RST) framework.
Our framework facilitates discourse parsing from scratch without requiring discourse segmentation as a prerequisite.
Our unified parsing model adopts a beam search to decode the best tree structure by searching through a space of high-scoring trees.
arXiv Detail & Related papers (2021-05-23T06:19:38Z) - Low-Resource Task-Oriented Semantic Parsing via Intrinsic Modeling [65.51280121472146]
We exploit what we intrinsically know about ontology labels to build efficient semantic parsing models.
Our model is highly efficient using a low-resource benchmark derived from TOPv2.
arXiv Detail & Related papers (2021-04-15T04:01:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.