Coreference Reasoning in Machine Reading Comprehension
- URL: http://arxiv.org/abs/2012.15573v1
- Date: Thu, 31 Dec 2020 12:18:41 GMT
- Title: Coreference Reasoning in Machine Reading Comprehension
- Authors: Mingzhu Wu, Nafise Sadat Moosavi, Dan Roth, Iryna Gurevych
- Abstract summary: We show that coreference reasoning in machine reading comprehension is a greater challenge than was earlier thought.
We propose a methodology for creating reading comprehension datasets that better reflect the challenges of coreference reasoning.
This allows us to show an improvement in the reasoning abilities of state-of-the-art models across various MRC datasets.
- Score: 100.75624364257429
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The ability to reason about multiple references to a given entity is
essential for natural language understanding and has been long studied in NLP.
In recent years, as the format of Question Answering (QA) became a standard for
machine reading comprehension (MRC), there have been data collection efforts,
e.g., Dasigi et al. (2019), that attempt to evaluate the ability of MRC models
to reason about coreference. However, as we show, coreference reasoning in MRC
is a greater challenge than was earlier thought; MRC datasets do not reflect
the natural distribution and, consequently, the challenges of coreference
reasoning. Specifically, success on these datasets does not reflect a model's
proficiency in coreference reasoning. We propose a methodology for creating
reading comprehension datasets that better reflect the challenges of
coreference reasoning and use it to show that state-of-the-art models still
struggle with these phenomena. Furthermore, we develop an effective way to use
naturally occurring coreference phenomena from annotated coreference resolution
datasets when training MRC models. This allows us to show an improvement in the
coreference reasoning abilities of state-of-the-art models across various MRC
datasets. We will release all the code and the resulting dataset at
https://github.com/UKPLab/coref-reasoning-in-qa.
Related papers
- Improve Vision Language Model Chain-of-thought Reasoning [86.83335752119741]
Chain-of-thought (CoT) reasoning in vision language models (VLMs) is crucial for improving interpretability and trustworthiness.
We show that training VLM on short answers does not generalize well to reasoning tasks that require more detailed responses.
arXiv Detail & Related papers (2024-10-21T17:00:06Z) - X-ray Made Simple: Radiology Report Generation and Evaluation with Layman's Terms [25.871814979179373]
Radiology Report Generation (RRG) has achieved significant progress with the advancements of multimodal generative models.
High performance on RRG with existing lexical-based metrics (e.g. BLEU) might be more of a mirage - a model can get a high BLEU only by learning the template of reports.
We un-intuitively approach this problem by proposing the Layman's RRG framework, a layman's terms-based dataset, evaluation and training framework.
arXiv Detail & Related papers (2024-06-25T19:52:01Z) - Evaluating Consistency and Reasoning Capabilities of Large Language Models [0.0]
Large Language Models (LLMs) are extensively used today across various sectors, including academia, research, business, and finance.
Despite their widespread adoption, these models often produce incorrect and misleading information, exhibiting a tendency to hallucinate.
This paper aims to evaluate and compare the consistency and reasoning capabilities of both public and proprietary LLMs.
arXiv Detail & Related papers (2024-04-25T10:03:14Z) - Lite Unified Modeling for Discriminative Reading Comprehension [68.39862736200045]
We propose a lightweight POS-Enhanced Iterative Co-Attention Network (POI-Net) to handle diverse discriminative MRC tasks synchronously.
Our lite unified design brings model significant improvement with both encoder and decoder components.
The evaluation results on four discriminative MRC benchmarks consistently indicate the general effectiveness and applicability of our model.
arXiv Detail & Related papers (2022-03-26T15:47:19Z) - Tracing Origins: Coref-aware Machine Reading Comprehension [43.352833140317486]
We imitated the human's reading process in connecting the anaphoric expressions and leverage the coreference information to enhance the word embeddings from the pre-trained model.
We demonstrated that the explicit incorporation of the coreference information in fine-tuning stage performed better than the incorporation of the coreference information in training a pre-trained language models.
arXiv Detail & Related papers (2021-10-15T09:28:35Z) - ExpMRC: Explainability Evaluation for Machine Reading Comprehension [42.483940360860096]
We propose a new benchmark called ExpMRC for evaluating the explainability of the Machine Reading systems.
We use state-of-the-art pre-trained language models to build baseline systems and adopt various unsupervised approaches to extract evidence without a human-annotated training set.
arXiv Detail & Related papers (2021-05-10T06:00:20Z) - Learning from Context or Names? An Empirical Study on Neural Relation
Extraction [112.06614505580501]
We study the effect of two main information sources in text: textual context and entity mentions (names)
We propose an entity-masked contrastive pre-training framework for relation extraction (RE)
Our framework can improve the effectiveness and robustness of neural models in different RE scenarios.
arXiv Detail & Related papers (2020-10-05T11:21:59Z) - A Brief Survey and Comparative Study of Recent Development of Pronoun
Coreference Resolution [55.39835612617972]
Pronoun Coreference Resolution (PCR) is the task of resolving pronominal expressions to all mentions they refer to.
As one important natural language understanding (NLU) component, pronoun resolution is crucial for many downstream tasks and still challenging for existing models.
We conduct extensive experiments to show that even though current models are achieving good performance on the standard evaluation set, they are still not ready to be used in real applications.
arXiv Detail & Related papers (2020-09-27T01:40:01Z) - ReClor: A Reading Comprehension Dataset Requiring Logical Reasoning [85.33459673197149]
We introduce a new Reading dataset requiring logical reasoning (ReClor) extracted from standardized graduate admission examinations.
In this paper, we propose to identify biased data points and separate them into EASY set and the rest as HARD set.
Empirical results show that state-of-the-art models have an outstanding ability to capture biases contained in the dataset with high accuracy on EASY set.
However, they struggle on HARD set with poor performance near that of random guess, indicating more research is needed to essentially enhance the logical reasoning ability of current models.
arXiv Detail & Related papers (2020-02-11T11:54:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.