Structural Modeling for Dialogue Disentanglement
- URL: http://arxiv.org/abs/2110.08018v1
- Date: Fri, 15 Oct 2021 11:28:43 GMT
- Title: Structural Modeling for Dialogue Disentanglement
- Authors: Xinbei Ma, Zhuosheng Zhang and Hai Zhao
- Abstract summary: Multi-party dialogue context Tangled multi-party dialogue context leads to challenges for dialogue reading comprehension.
This work designs a novel model to disentangle multi-party history into threads, by taking dialogue structure features into account.
- Score: 43.352833140317486
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Tangled multi-party dialogue context leads to challenges for dialogue reading
comprehension, where multiple dialogue threads flow simultaneously within the
same dialogue history, thus increasing difficulties in understanding a dialogue
history for both human and machine. Dialogue disentanglement aims to clarify
conversation threads in a multi-party dialogue history, thus reducing the
difficulty of comprehending the long disordered dialogue passage. Existing
studies commonly focus on utterance encoding with carefully designed feature
engineering-based methods but pay inadequate attention to dialogue structure.
This work designs a novel model to disentangle multi-party history into
threads, by taking dialogue structure features into account. Specifically,
based on the fact that dialogues are constructed through successive
participation of speakers and interactions between users of interest, we
extract clues of speaker property and reference of users to model the structure
of a long dialogue record. The novel method is evaluated on the Ubuntu IRC
dataset and shows state-of-the-art experimental results in dialogue
disentanglement.
Related papers
- A Static and Dynamic Attention Framework for Multi Turn Dialogue Generation [37.79563028123686]
In open domain multi turn dialogue generation, it is essential to modeling the contextual semantics of the dialogue history.
Previous research had verified the effectiveness of the hierarchical recurrent encoder-decoder framework on open domain multi turn dialogue generation.
We propose a static and dynamic attention-based approach to model the dialogue history and then generate open domain multi turn dialogue responses.
arXiv Detail & Related papers (2024-10-28T06:05:34Z) - Multi-turn Dialogue Comprehension from a Topic-aware Perspective [70.37126956655985]
This paper proposes to model multi-turn dialogues from a topic-aware perspective.
We use a dialogue segmentation algorithm to split a dialogue passage into topic-concentrated fragments in an unsupervised way.
We also present a novel model, Topic-Aware Dual-Attention Matching (TADAM) Network, which takes topic segments as processing elements.
arXiv Detail & Related papers (2023-09-18T11:03:55Z) - STRUDEL: Structured Dialogue Summarization for Dialogue Comprehension [42.57581945778631]
Abstractive dialogue summarization has long been viewed as an important standalone task in natural language processing.
We propose a novel type of dialogue summarization task - STRUctured DiaLoguE Summarization.
We show that our STRUDEL dialogue comprehension model can significantly improve the dialogue comprehension performance of transformer encoder language models.
arXiv Detail & Related papers (2022-12-24T04:39:54Z) - A Benchmark for Understanding and Generating Dialogue between Characters
in Stories [75.29466820496913]
We present the first study to explore whether machines can understand and generate dialogue in stories.
We propose two new tasks including Masked Dialogue Generation and Dialogue Speaker Recognition.
We show the difficulty of the proposed tasks by testing existing models with automatic and manual evaluation on DialStory.
arXiv Detail & Related papers (2022-09-18T10:19:04Z) - Advances in Multi-turn Dialogue Comprehension: A Survey [51.215629336320305]
Training machines to understand natural language and interact with humans is an elusive and essential task of artificial intelligence.
This paper reviews the previous methods from the technical perspective of dialogue modeling for the dialogue comprehension task.
In addition, we categorize dialogue-related pre-training techniques which are employed to enhance PrLMs in dialogue scenarios.
arXiv Detail & Related papers (2021-10-11T03:52:37Z) - DialogLM: Pre-trained Model for Long Dialogue Understanding and
Summarization [19.918194137007653]
We present a pre-training framework for long dialogue understanding and summarization.
Considering the nature of long conversations, we propose a window-based denoising approach for generative pre-training.
We conduct extensive experiments on five datasets of long dialogues, covering tasks of dialogue summarization, abstractive question answering and topic segmentation.
arXiv Detail & Related papers (2021-09-06T13:55:03Z) - Advances in Multi-turn Dialogue Comprehension: A Survey [51.215629336320305]
We review the previous methods from the perspective of dialogue modeling.
We discuss three typical patterns of dialogue modeling that are widely-used in dialogue comprehension tasks.
arXiv Detail & Related papers (2021-03-04T15:50:17Z) - Rethinking Dialogue State Tracking with Reasoning [76.0991910623001]
This paper proposes to track dialogue states gradually with reasoning over dialogue turns with the help of the back-end data.
Empirical results demonstrate that our method significantly outperforms the state-of-the-art methods by 38.6% in terms of joint belief accuracy for MultiWOZ 2.1.
arXiv Detail & Related papers (2020-05-27T02:05:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.