Semantic Role Labeling Guided Multi-turn Dialogue ReWriter
- URL: http://arxiv.org/abs/2010.01417v1
- Date: Sat, 3 Oct 2020 19:50:04 GMT
- Title: Semantic Role Labeling Guided Multi-turn Dialogue ReWriter
- Authors: Kun Xu and Haochen Tan and Linfeng Song and Han Wu and Haisong Zhang
and Linqi Song and Dong Yu
- Abstract summary: We propose to use semantic role labeling (SRL) to highlight the core semantic information of who did what to whom.
Experiments show that this information significantly improves a RoBERTa-based model that already outperforms previous state-of-the-art systems.
- Score: 63.07073750355096
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: For multi-turn dialogue rewriting, the capacity of effectively modeling the
linguistic knowledge in dialog context and getting rid of the noises is
essential to improve its performance. Existing attentive models attend to all
words without prior focus, which results in inaccurate concentration on some
dispensable words. In this paper, we propose to use semantic role labeling
(SRL), which highlights the core semantic information of who did what to whom,
to provide additional guidance for the rewriter model. Experiments show that
this information significantly improves a RoBERTa-based model that already
outperforms previous state-of-the-art systems.
Related papers
- Integrating Self-supervised Speech Model with Pseudo Word-level Targets
from Visually-grounded Speech Model [57.78191634042409]
We propose Pseudo-Word HuBERT (PW-HuBERT), a framework that integrates pseudo word-level targets into the training process.
Our experimental results on four spoken language understanding (SLU) benchmarks suggest the superiority of our model in capturing semantic information.
arXiv Detail & Related papers (2024-02-08T16:55:21Z) - SPECTRUM: Speaker-Enhanced Pre-Training for Long Dialogue Summarization [48.284512017469524]
Multi-turn dialogues are characterized by their extended length and the presence of turn-taking conversations.
Traditional language models often overlook the distinct features of these dialogues by treating them as regular text.
We propose a speaker-enhanced pre-training method for long dialogue summarization.
arXiv Detail & Related papers (2024-01-31T04:50:00Z) - Improved Contextual Recognition In Automatic Speech Recognition Systems
By Semantic Lattice Rescoring [4.819085609772069]
We propose a novel approach for enhancing contextual recognition within ASR systems via semantic lattice processing.
Our solution consists of using Hidden Markov Models and Gaussian Mixture Models (HMM-GMM) along with Deep Neural Networks (DNN) models for better accuracy.
We demonstrate the effectiveness of our proposed framework on the LibriSpeech dataset with empirical analyses.
arXiv Detail & Related papers (2023-10-14T23:16:05Z) - Pre-training Multi-party Dialogue Models with Latent Discourse Inference [85.9683181507206]
We pre-train a model that understands the discourse structure of multi-party dialogues, namely, to whom each utterance is replying.
To fully utilize the unlabeled data, we propose to treat the discourse structures as latent variables, then jointly infer them and pre-train the discourse-aware model.
arXiv Detail & Related papers (2023-05-24T14:06:27Z) - Conversational Semantic Role Labeling with Predicate-Oriented Latent
Graph [40.43625257213158]
We propose to automatically induce a predicate-oriented latent graph (POLar) with a predicate-centered Gaussian mechanism.
The POLar structure is then dynamically pruned and refined so as to best fit the task need.
We additionally introduce an effective dialogue-level pre-trained language model, CoDiaBERT, for better supporting multiple utterance sentences.
arXiv Detail & Related papers (2022-10-06T16:42:00Z) - Utterance Rewriting with Contrastive Learning in Multi-turn Dialogue [22.103162555263143]
We introduce contrastive learning and multi-task learning to jointly model the problem.
Our proposed model achieves state-of-the-art performance on several public datasets.
arXiv Detail & Related papers (2022-03-22T10:13:27Z) - CSAGN: Conversational Structure Aware Graph Network for Conversational
Semantic Role Labeling [27.528361001332264]
We present a simple and effective architecture for CSRL which aims to address this problem.
Our model is based on a conversational structure-aware graph network which explicitly encodes the speaker dependent information.
arXiv Detail & Related papers (2021-09-23T07:47:28Z) - Enhancing Dialogue Generation via Multi-Level Contrastive Learning [57.005432249952406]
We propose a multi-level contrastive learning paradigm to model the fine-grained quality of the responses with respect to the query.
A Rank-aware (RC) network is designed to construct the multi-level contrastive optimization objectives.
We build a Knowledge Inference (KI) component to capture the keyword knowledge from the reference during training and exploit such information to encourage the generation of informative words.
arXiv Detail & Related papers (2020-09-19T02:41:04Z) - Filling the Gap of Utterance-aware and Speaker-aware Representation for
Multi-turn Dialogue [76.88174667929665]
A multi-turn dialogue is composed of multiple utterances from two or more different speaker roles.
In the existing retrieval-based multi-turn dialogue modeling, the pre-trained language models (PrLMs) as encoder represent the dialogues coarsely.
We propose a novel model to fill such a gap by modeling the effective utterance-aware and speaker-aware representations entailed in a dialogue history.
arXiv Detail & Related papers (2020-09-14T15:07:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.