Multi-tasking Dialogue Comprehension with Discourse Parsing
- URL: http://arxiv.org/abs/2110.03269v1
- Date: Thu, 7 Oct 2021 08:51:49 GMT
- Title: Multi-tasking Dialogue Comprehension with Discourse Parsing
- Authors: Yuchen He, Zhuosheng Zhang, Hai Zhao
- Abstract summary: We propose the first multi-task model for jointly performing QA and discourse parsing (DP) on the multi-party dialogue MRC task.
Our results indicate that training with complementary tasks indeed benefits not only QA task, but also DP task itself.
- Score: 43.352833140317486
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Multi-party dialogue machine reading comprehension (MRC) raises an even more
challenging understanding goal on dialogue with more than two involved
speakers, compared with the traditional plain passage style MRC. To accurately
perform the question-answering (QA) task according to such multi-party
dialogue, models have to handle fundamentally different discourse relationships
from common non-dialogue plain text, where discourse relations are supposed to
connect two far apart utterances in a linguistics-motivated way.To further
explore the role of such unusual discourse structure on the correlated QA task
in terms of MRC, we propose the first multi-task model for jointly performing
QA and discourse parsing (DP) on the multi-party dialogue MRC task. Our
proposed model is evaluated on the latest benchmark Molweni, whose results
indicate that training with complementary tasks indeed benefits not only QA
task, but also DP task itself. We further find that the joint model is
distinctly stronger when handling longer dialogues which again verifies the
necessity of DP in the related MRC.
Related papers
- Self-Explanation Prompting Improves Dialogue Understanding in Large
Language Models [52.24756457516834]
We propose a novel "Self-Explanation" prompting strategy to enhance the comprehension abilities of Large Language Models (LLMs)
This task-agnostic approach requires the model to analyze each dialogue utterance before task execution, thereby improving performance across various dialogue-centric tasks.
Experimental results from six benchmark datasets confirm that our method consistently outperforms other zero-shot prompts and matches or exceeds the efficacy of few-shot prompts.
arXiv Detail & Related papers (2023-09-22T15:41:34Z) - Coreference-aware Double-channel Attention Network for Multi-party
Dialogue Reading Comprehension [7.353227696624305]
We tackle Multi-party Dialogue Reading (abbr., MDRC)
MDRC stands for an extractive reading comprehension task grounded on a batch of dialogues among multiple interlocutors.
We propose a coreference-aware attention modeling method to strengthen the reasoning ability.
arXiv Detail & Related papers (2023-05-15T05:01:29Z) - ChatGPT Evaluation on Sentence Level Relations: A Focus on Temporal,
Causal, and Discourse Relations [52.26802326949116]
We quantitatively evaluate the performance of ChatGPT, an interactive large language model, on inter-sentential relations.
ChatGPT exhibits exceptional proficiency in detecting and reasoning about causal relations.
It is capable of identifying the majority of discourse relations with existing explicit discourse connectives, but the implicit discourse relation remains a formidable challenge.
arXiv Detail & Related papers (2023-04-28T13:14:36Z) - TOD-DA: Towards Boosting the Robustness of Task-oriented Dialogue
Modeling on Spoken Conversations [24.245354500835465]
We propose a novel model-agnostic data augmentation paradigm to boost the robustness of task-oriented dialogue modeling on spoken conversations.
Our approach ranked first in both tasks of DSTC10 Track2, a benchmark for task-oriented dialogue modeling on spoken conversations.
arXiv Detail & Related papers (2021-12-23T10:04:25Z) - Self- and Pseudo-self-supervised Prediction of Speaker and Key-utterance
for Multi-party Dialogue Reading Comprehension [46.69961067676279]
Multi-party dialogue machine reading comprehension (MRC) brings tremendous challenge since it involves multiple speakers at one dialogue.
Previous models focus on how to incorporate speaker information flows using complex graph-based modules.
In this paper, we design two labour-free self- and pseudo-self-supervised prediction tasks on speaker and key-utterance to implicitly model the speaker information flows.
arXiv Detail & Related papers (2021-09-08T16:51:41Z) - MPC-BERT: A Pre-Trained Language Model for Multi-Party Conversation
Understanding [58.95156916558384]
We present MPC-BERT, a pre-trained model for MPC understanding.
We evaluate MPC-BERT on three downstream tasks including addressee recognition, speaker identification and response selection.
arXiv Detail & Related papers (2021-06-03T01:49:12Z) - Structural Pre-training for Dialogue Comprehension [51.215629336320305]
We present SPIDER, Structural Pre-traIned DialoguE Reader, to capture dialogue exclusive features.
To simulate the dialogue-like features, we propose two training objectives in addition to the original LM objectives.
Experimental results on widely used dialogue benchmarks verify the effectiveness of the newly introduced self-supervised tasks.
arXiv Detail & Related papers (2021-05-23T15:16:54Z) - Matching Questions and Answers in Dialogues from Online Forums [12.64602629459043]
Matching question-answer relations between two turns in conversations is not only the first step in analyzing dialogue structures, but also valuable for training dialogue systems.
This paper presents a QA matching model considering both distance information and dialogue history by two simultaneous attention mechanisms called mutual attention.
arXiv Detail & Related papers (2020-05-19T08:18:52Z) - Dialogue-Based Relation Extraction [53.2896545819799]
We present the first human-annotated dialogue-based relation extraction (RE) dataset DialogRE.
We argue that speaker-related information plays a critical role in the proposed task, based on an analysis of similarities and differences between dialogue-based and traditional RE tasks.
Experimental results demonstrate that a speaker-aware extension on the best-performing model leads to gains in both the standard and conversational evaluation settings.
arXiv Detail & Related papers (2020-04-17T03:51:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.