EviDR: Evidence-Emphasized Discrete Reasoning for Reasoning Machine
Reading Comprehension
- URL: http://arxiv.org/abs/2108.07994v1
- Date: Wed, 18 Aug 2021 06:49:58 GMT
- Title: EviDR: Evidence-Emphasized Discrete Reasoning for Reasoning Machine
Reading Comprehension
- Authors: Yongwei Zhou, Junwei Bao, Haipeng Sun, Jiahui Liang, Youzheng Wu,
Xiaodong He, Bowen Zhou, and Tiejun Zhao
- Abstract summary: Reasoning machine reading comprehension (R-MRC) aims to answer complex questions that require discrete reasoning based on text.
Previous end-to-end methods that achieve state-of-the-art performance rarely solve the problem by paying enough emphasis on the modeling of evidence.
We propose an evidence-emphasized discrete reasoning approach (EviDR), in which sentence and clause level evidence is first detected based on distant supervision.
- Score: 39.970232108247394
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Reasoning machine reading comprehension (R-MRC) aims to answer complex
questions that require discrete reasoning based on text. To support discrete
reasoning, evidence, typically the concise textual fragments that describe
question-related facts, including topic entities and attribute values, are
crucial clues from question to answer. However, previous end-to-end methods
that achieve state-of-the-art performance rarely solve the problem by paying
enough emphasis on the modeling of evidence, missing the opportunity to further
improve the model's reasoning ability for R-MRC. To alleviate the above issue,
in this paper, we propose an evidence-emphasized discrete reasoning approach
(EviDR), in which sentence and clause level evidence is first detected based on
distant supervision, and then used to drive a reasoning module implemented with
a relational heterogeneous graph convolutional network to derive answers.
Extensive experiments are conducted on DROP (discrete reasoning over
paragraphs) dataset, and the results demonstrate the effectiveness of our
proposed approach. In addition, qualitative analysis verifies the capability of
the proposed evidence-emphasized discrete reasoning for R-MRC.
Related papers
- Eliciting Critical Reasoning in Retrieval-Augmented Language Models via Contrastive Explanations [4.697267141773321]
Retrieval-augmented generation (RAG) has emerged as a critical mechanism in contemporary NLP to support Large Language Models (LLMs) in systematically accessing richer factual context.
Recent studies have shown that LLMs still struggle to critically analyse RAG-based in-context information, a limitation that may lead to incorrect inferences and hallucinations.
In this paper, we investigate how to elicit critical reasoning in RAG via contrastive explanations.
arXiv Detail & Related papers (2024-10-30T10:11:53Z) - Inducing Causal Structure for Abstractive Text Summarization [76.1000380429553]
We introduce a Structural Causal Model (SCM) to induce the underlying causal structure of the summarization data.
We propose a Causality Inspired Sequence-to-Sequence model (CI-Seq2Seq) to learn the causal representations that can mimic the causal factors.
Experimental results on two widely used text summarization datasets demonstrate the advantages of our approach.
arXiv Detail & Related papers (2023-08-24T16:06:36Z) - Towards Trustworthy Explanation: On Causal Rationalization [9.48539398357156]
We propose a new model of rationalization based on two causal desiderata, non-spuriousness and efficiency.
The superior performance of the proposed causal rationalization is demonstrated on real-world review and medical datasets.
arXiv Detail & Related papers (2023-06-25T03:34:06Z) - Advancing Counterfactual Inference through Nonlinear Quantile Regression [77.28323341329461]
We propose a framework for efficient and effective counterfactual inference implemented with neural networks.
The proposed approach enhances the capacity to generalize estimated counterfactual outcomes to unseen data.
Empirical results conducted on multiple datasets offer compelling support for our theoretical assertions.
arXiv Detail & Related papers (2023-06-09T08:30:51Z) - Learning a Structural Causal Model for Intuition Reasoning in
Conversation [20.243323155177766]
Reasoning, a crucial aspect of NLP research, has not been adequately addressed by prevailing models.
We develop a conversation cognitive model ( CCM) that explains how each utterance receives and activates channels of information.
By leveraging variational inference, it explores substitutes for implicit causes, addresses the issue of their unobservability, and reconstructs the causal representations of utterances through the evidence lower bounds.
arXiv Detail & Related papers (2023-05-28T13:54:09Z) - Logical Satisfiability of Counterfactuals for Faithful Explanations in
NLI [60.142926537264714]
We introduce the methodology of Faithfulness-through-Counterfactuals.
It generates a counterfactual hypothesis based on the logical predicates expressed in the explanation.
It then evaluates if the model's prediction on the counterfactual is consistent with that expressed logic.
arXiv Detail & Related papers (2022-05-25T03:40:59Z) - Reasoning over Hybrid Chain for Table-and-Text Open Domain QA [69.8436986668218]
We propose a ChAin-centric Reasoning and Pre-training framework (CARP)
CARP utilizes hybrid chain to model the explicit intermediate reasoning process across table and text for question answering.
We also propose a novel chain-centric pre-training method, to enhance the pre-trained model in identifying the cross-modality reasoning process.
arXiv Detail & Related papers (2022-01-15T16:11:55Z) - Discrete Reasoning Templates for Natural Language Understanding [79.07883990966077]
We present an approach that reasons about complex questions by decomposing them to simpler subquestions.
We derive the final answer according to instructions in a predefined reasoning template.
We show that our approach is competitive with the state-of-the-art while being interpretable and requires little supervision.
arXiv Detail & Related papers (2021-04-05T18:56:56Z) - Multi-hop Inference for Question-driven Summarization [39.08269647808958]
We propose a novel question-driven abstractive summarization method, Multi-hop Selective Generator (MSG)
MSG incorporates multi-hop reasoning into question-driven summarization and, meanwhile, provide justifications for the generated summaries.
Experimental results show that the proposed method consistently outperforms state-of-the-art methods on two non-factoid QA datasets.
arXiv Detail & Related papers (2020-10-08T02:36:39Z) - Context Modeling with Evidence Filter for Multiple Choice Question
Answering [18.154792554957595]
Multiple-Choice Question Answering (MCQA) is a challenging task in machine reading comprehension.
The main challenge is to extract "evidence" from the given context that supports the correct answer.
Existing work tackles this problem by annotated evidence or distant supervision with rules which overly rely on human efforts.
We propose a simple yet effective approach termed evidence filtering to model the relationships between the encoded contexts.
arXiv Detail & Related papers (2020-10-06T11:53:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.