Coreference-aware Double-channel Attention Network for Multi-party
Dialogue Reading Comprehension
- URL: http://arxiv.org/abs/2305.08348v2
- Date: Mon, 22 May 2023 04:42:40 GMT
- Title: Coreference-aware Double-channel Attention Network for Multi-party
Dialogue Reading Comprehension
- Authors: Yanling Li, Bowei Zou, Yifan Fan, Mengxing Dong, Yu Hong
- Abstract summary: We tackle Multi-party Dialogue Reading (abbr., MDRC)
MDRC stands for an extractive reading comprehension task grounded on a batch of dialogues among multiple interlocutors.
We propose a coreference-aware attention modeling method to strengthen the reasoning ability.
- Score: 7.353227696624305
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We tackle Multi-party Dialogue Reading Comprehension (abbr., MDRC). MDRC
stands for an extractive reading comprehension task grounded on a batch of
dialogues among multiple interlocutors. It is challenging due to the
requirement of understanding cross-utterance contexts and relationships in a
multi-turn multi-party conversation. Previous studies have made great efforts
on the utterance profiling of a single interlocutor and graph-based interaction
modeling. The corresponding solutions contribute to the answer-oriented
reasoning on a series of well-organized and thread-aware conversational
contexts. However, the current MDRC models still suffer from two bottlenecks.
On the one hand, a pronoun like "it" most probably produces multi-skip
reasoning throughout the utterances of different interlocutors. On the other
hand, an MDRC encoder is potentially puzzled by fuzzy features, i.e., the
mixture of inner linguistic features in utterances and external interactive
features among utterances. To overcome the bottlenecks, we propose a
coreference-aware attention modeling method to strengthen the reasoning
ability. In addition, we construct a two-channel encoding network. It
separately encodes utterance profiles and interactive relationships, so as to
relieve the confusion among heterogeneous features. We experiment on the
benchmark corpora Molweni and FriendsQA. Experimental results demonstrate that
our approach yields substantial improvements on both corpora, compared to the
fine-tuned BERT and ELECTRA baselines. The maximum performance gain is about
2.5\% F1-score. Besides, our MDRC models outperform the state-of-the-art in
most cases.
Related papers
- Explaining Interactions Between Text Spans [50.70253702800355]
Reasoning over spans of tokens from different parts of the input is essential for natural language understanding.
We introduce SpanEx, a dataset of human span interaction explanations for two NLU tasks: NLI and FC.
We then investigate the decision-making processes of multiple fine-tuned large language models in terms of the employed connections between spans.
arXiv Detail & Related papers (2023-10-20T13:52:37Z) - Dynamic Multi-Scale Context Aggregation for Conversational Aspect-Based
Sentiment Quadruple Analysis [4.768182075837568]
DiaASQ aims to extract the quadruple of target-aspect-opinion-sentiment within a dialogue.
Existing work independently encodes each utterance, thereby struggling to capture long-range conversational context.
We propose a novel Dynamic Multi-scale Context Aggregation network (DMCA) to address the challenges.
arXiv Detail & Related papers (2023-09-27T08:17:28Z) - Multi-turn Dialogue Comprehension from a Topic-aware Perspective [70.37126956655985]
This paper proposes to model multi-turn dialogues from a topic-aware perspective.
We use a dialogue segmentation algorithm to split a dialogue passage into topic-concentrated fragments in an unsupervised way.
We also present a novel model, Topic-Aware Dual-Attention Matching (TADAM) Network, which takes topic segments as processing elements.
arXiv Detail & Related papers (2023-09-18T11:03:55Z) - DiPlomat: A Dialogue Dataset for Situated Pragmatic Reasoning [89.92601337474954]
Pragmatic reasoning plays a pivotal role in deciphering implicit meanings that frequently arise in real-life conversations.
We introduce a novel challenge, DiPlomat, aiming at benchmarking machines' capabilities on pragmatic reasoning and situated conversational understanding.
arXiv Detail & Related papers (2023-06-15T10:41:23Z) - Pre-training Multi-party Dialogue Models with Latent Discourse Inference [85.9683181507206]
We pre-train a model that understands the discourse structure of multi-party dialogues, namely, to whom each utterance is replying.
To fully utilize the unlabeled data, we propose to treat the discourse structures as latent variables, then jointly infer them and pre-train the discourse-aware model.
arXiv Detail & Related papers (2023-05-24T14:06:27Z) - IRRGN: An Implicit Relational Reasoning Graph Network for Multi-turn
Response Selection [4.471148909362883]
Implicit Reasoning to Graph Network aims to implicitly extract between utterances, as well as utterances and options.
Model surpasses human performance for the first time on the MuTual dataset.
arXiv Detail & Related papers (2022-12-01T13:17:25Z) - Findings on Conversation Disentanglement [28.874162427052905]
We build a learning model that learns utterance-to-utterance and utterance-to-thread classification.
Experiments on the Ubuntu IRC dataset show that this approach has the potential to outperform the conventional greedy approach.
arXiv Detail & Related papers (2021-12-10T05:54:48Z) - Multi-tasking Dialogue Comprehension with Discourse Parsing [43.352833140317486]
We propose the first multi-task model for jointly performing QA and discourse parsing (DP) on the multi-party dialogue MRC task.
Our results indicate that training with complementary tasks indeed benefits not only QA task, but also DP task itself.
arXiv Detail & Related papers (2021-10-07T08:51:49Z) - Filling the Gap of Utterance-aware and Speaker-aware Representation for
Multi-turn Dialogue [76.88174667929665]
A multi-turn dialogue is composed of multiple utterances from two or more different speaker roles.
In the existing retrieval-based multi-turn dialogue modeling, the pre-trained language models (PrLMs) as encoder represent the dialogues coarsely.
We propose a novel model to fill such a gap by modeling the effective utterance-aware and speaker-aware representations entailed in a dialogue history.
arXiv Detail & Related papers (2020-09-14T15:07:19Z) - Multi-Stage Conversational Passage Retrieval: An Approach to Fusing Term
Importance Estimation and Neural Query Rewriting [56.268862325167575]
We tackle conversational passage retrieval (ConvPR) with query reformulation integrated into a multi-stage ad-hoc IR system.
We propose two conversational query reformulation (CQR) methods: (1) term importance estimation and (2) neural query rewriting.
For the former, we expand conversational queries using important terms extracted from the conversational context with frequency-based signals.
For the latter, we reformulate conversational queries into natural, standalone, human-understandable queries with a pretrained sequence-tosequence model.
arXiv Detail & Related papers (2020-05-05T14:30:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.