Revisiting Conversation Discourse for Dialogue Disentanglement
- URL: http://arxiv.org/abs/2306.03975v2
- Date: Sat, 10 Jun 2023 06:42:03 GMT
- Title: Revisiting Conversation Discourse for Dialogue Disentanglement
- Authors: Bobo Li, Hao Fei, Fei Li, Shengqiong Wu, Lizi Liao, Yinwei Wei,
Tat-Seng Chua, Donghong Ji
- Abstract summary: We propose enhancing dialogue disentanglement by taking full advantage of the dialogue discourse characteristics.
We develop a structure-aware framework to integrate the rich structural features for better modeling the conversational semantic context.
Our work has great potential to facilitate broader multi-party multi-thread dialogue applications.
- Score: 88.3386821205896
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Dialogue disentanglement aims to detach the chronologically ordered
utterances into several independent sessions. Conversation utterances are
essentially organized and described by the underlying discourse, and thus
dialogue disentanglement requires the full understanding and harnessing of the
intrinsic discourse attribute. In this paper, we propose enhancing dialogue
disentanglement by taking full advantage of the dialogue discourse
characteristics. First of all, in feature encoding stage, we construct the
heterogeneous graph representations to model the various dialogue-specific
discourse structural features, including the static speaker-role structures
(i.e., speaker-utterance and speaker-mentioning structure) and the dynamic
contextual structures (i.e., the utterance-distance and partial-replying
structure). We then develop a structure-aware framework to integrate the rich
structural features for better modeling the conversational semantic context.
Second, in model learning stage, we perform optimization with a hierarchical
ranking loss mechanism, which groups dialogue utterances into different
discourse levels and carries training covering pair-wise and session-wise
levels hierarchically. Third, in inference stage, we devise an easy-first
decoding algorithm, which performs utterance pairing under the easy-to-hard
manner with a global context, breaking the constraint of traditional sequential
decoding order. On two benchmark datasets, our overall system achieves new
state-of-the-art performances on all evaluations. In-depth analyses further
demonstrate the efficacy of each proposed idea and also reveal how our methods
help advance the task. Our work has great potential to facilitate broader
multi-party multi-thread dialogue applications.
Related papers
- Unsupervised Mutual Learning of Dialogue Discourse Parsing and Topic Segmentation [38.956438905614256]
rhetorical structure and topic structure are mostly modeled separately or with one assisting the other in the prior work.
We propose an unsupervised mutual learning framework of two structures leveraging the global and local connections between them.
We also incorporate rhetorical structures into the topic structure through a graph neural network model to ensure local coherence consistency.
arXiv Detail & Related papers (2024-05-30T08:10:50Z) - With a Little Help from my (Linguistic) Friends: Topic Segmentation of
Multi-party Casual Conversations [0.565395466029518]
This paper tries to reach a comparable level of accuracy as recent machine learning based topic segmentation models.
The features we identify as meaningful for this task help us understand better the topical structure of a conversation.
arXiv Detail & Related papers (2024-02-05T09:48:07Z) - Channel-aware Decoupling Network for Multi-turn Dialogue Comprehension [81.47133615169203]
We propose compositional learning for holistic interaction across utterances beyond the sequential contextualization from PrLMs.
We employ domain-adaptive training strategies to help the model adapt to the dialogue domains.
Experimental results show that our method substantially boosts the strong PrLM baselines in four public benchmark datasets.
arXiv Detail & Related papers (2023-01-10T13:18:25Z) - Unsupervised Learning of Hierarchical Conversation Structure [50.29889385593043]
Goal-oriented conversations often have meaningful sub-dialogue structure, but it can be highly domain-dependent.
This work introduces an unsupervised approach to learning hierarchical conversation structure, including turn and sub-dialogue segment labels.
The decoded structure is shown to be useful in enhancing neural models of language for three conversation-level understanding tasks.
arXiv Detail & Related papers (2022-05-24T17:52:34Z) - Who says like a style of Vitamin: Towards Syntax-Aware
DialogueSummarization using Multi-task Learning [2.251583286448503]
We focus on the association between utterances from individual speakers and unique syntactic structures.
Speakers have unique textual styles that can contain linguistic information, such as voiceprint.
We employ multi-task learning of both syntax-aware information and dialogue summarization.
arXiv Detail & Related papers (2021-09-29T05:30:39Z) - Structured Attention for Unsupervised Dialogue Structure Induction [110.12561786644122]
We propose to incorporate structured attention layers into a Variational Recurrent Neural Network (VRNN) model with discrete latent states to learn dialogue structure in an unsupervised fashion.
Compared to a vanilla VRNN, structured attention enables a model to focus on different parts of the source sentence embeddings while enforcing a structural inductive bias.
arXiv Detail & Related papers (2020-09-17T23:07:03Z) - Filling the Gap of Utterance-aware and Speaker-aware Representation for
Multi-turn Dialogue [76.88174667929665]
A multi-turn dialogue is composed of multiple utterances from two or more different speaker roles.
In the existing retrieval-based multi-turn dialogue modeling, the pre-trained language models (PrLMs) as encoder represent the dialogues coarsely.
We propose a novel model to fill such a gap by modeling the effective utterance-aware and speaker-aware representations entailed in a dialogue history.
arXiv Detail & Related papers (2020-09-14T15:07:19Z) - Masking Orchestration: Multi-task Pretraining for Multi-role Dialogue
Representation Learning [50.5572111079898]
Multi-role dialogue understanding comprises a wide range of diverse tasks such as question answering, act classification, dialogue summarization etc.
While dialogue corpora are abundantly available, labeled data, for specific learning tasks, can be highly scarce and expensive.
In this work, we investigate dialogue context representation learning with various types unsupervised pretraining tasks.
arXiv Detail & Related papers (2020-02-27T04:36:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.