Document Modeling with Graph Attention Networks for Multi-grained
Machine Reading Comprehension
- URL: http://arxiv.org/abs/2005.05806v2
- Date: Wed, 13 May 2020 08:44:37 GMT
- Title: Document Modeling with Graph Attention Networks for Multi-grained
Machine Reading Comprehension
- Authors: Bo Zheng, Haoyang Wen, Yaobo Liang, Nan Duan, Wanxiang Che, Daxin
Jiang, Ming Zhou and Ting Liu
- Abstract summary: Natural Questions is a new challenging machine reading comprehension benchmark.
It has two-grained answers, which are a long answer (typically a paragraph) and a short answer (one or more entities inside the long answer)
Existing methods treat these two sub-tasks individually during training while ignoring their dependencies.
We present a novel multi-grained machine reading comprehension framework that focuses on modeling documents at their hierarchical nature.
- Score: 127.3341842928421
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Natural Questions is a new challenging machine reading comprehension
benchmark with two-grained answers, which are a long answer (typically a
paragraph) and a short answer (one or more entities inside the long answer).
Despite the effectiveness of existing methods on this benchmark, they treat
these two sub-tasks individually during training while ignoring their
dependencies. To address this issue, we present a novel multi-grained machine
reading comprehension framework that focuses on modeling documents at their
hierarchical nature, which are different levels of granularity: documents,
paragraphs, sentences, and tokens. We utilize graph attention networks to
obtain different levels of representations so that they can be learned
simultaneously. The long and short answers can be extracted from
paragraph-level representation and token-level representation, respectively. In
this way, we can model the dependencies between the two-grained answers to
provide evidence for each other. We jointly train the two sub-tasks, and our
experiments show that our approach significantly outperforms previous systems
at both long and short answer criteria.
Related papers
- Recurrent Attention Networks for Long-text Modeling [14.710722261441822]
This paper proposes a novel long-document encoding model, Recurrent Attention Network (RAN), to enable the recurrent operation of self-attention.
RAN is capable of extracting global semantics in both token-level and document-level representations, making it inherently compatible with both sequential and classification tasks.
arXiv Detail & Related papers (2023-06-12T03:28:33Z) - UniKGQA: Unified Retrieval and Reasoning for Solving Multi-hop Question
Answering Over Knowledge Graph [89.98762327725112]
Multi-hop Question Answering over Knowledge Graph(KGQA) aims to find the answer entities that are multiple hops away from the topic entities mentioned in a natural language question.
We propose UniKGQA, a novel approach for multi-hop KGQA task, by unifying retrieval and reasoning in both model architecture and parameter learning.
arXiv Detail & Related papers (2022-12-02T04:08:09Z) - Summarization with Graphical Elements [55.5913491389047]
We propose a new task: summarization with graphical elements.
We collect a high quality human labeled dataset to support research into the task.
arXiv Detail & Related papers (2022-04-15T17:16:41Z) - Deep Understanding based Multi-Document Machine Reading Comprehension [22.319892892352414]
We propose a deep understanding based model for multi-document machine reading comprehension.
It has three cascaded deep understanding modules which are designed to understand the accurate semantic meaning of words.
We evaluate our model on two large scale benchmark datasets, namely TriviaQA Web and DuReader.
arXiv Detail & Related papers (2022-02-25T12:56:02Z) - MGA-VQA: Multi-Granularity Alignment for Visual Question Answering [75.55108621064726]
Learning to answer visual questions is a challenging task since the multi-modal inputs are within two feature spaces.
We propose Multi-Granularity Alignment architecture for Visual Question Answering task (MGA-VQA)
Our model splits alignment into different levels to achieve learning better correlations without needing additional data and annotations.
arXiv Detail & Related papers (2022-01-25T22:30:54Z) - MoCA: Incorporating Multi-stage Domain Pretraining and Cross-guided
Multimodal Attention for Textbook Question Answering [7.367945534481411]
We propose a novel model named MoCA, which incorporates multi-stage domain pretraining and multimodal cross attention for the Textbook Question Answering task.
The experimental results show the superiority of our model, which outperforms the state-of-the-art methods by 2.21% and 2.43% for validation and test split respectively.
arXiv Detail & Related papers (2021-12-06T07:58:53Z) - Cooperative Learning of Zero-Shot Machine Reading Comprehension [9.868221447090855]
We propose a cooperative, self-play learning model for question generation and answering.
We can train question generation and answering models on any textual corpora without annotation.
Our model outperforms the state-of-the-art pretrained language models on standard question answering benchmarks.
arXiv Detail & Related papers (2021-03-12T18:22:28Z) - Tradeoffs in Sentence Selection Techniques for Open-Domain Question
Answering [54.541952928070344]
We describe two groups of models for sentence selection: QA-based approaches, which run a full-fledged QA system to identify answer candidates, and retrieval-based models, which find parts of each passage specifically related to each question.
We show that very lightweight QA models can do well at this task, but retrieval-based models are faster still.
arXiv Detail & Related papers (2020-09-18T23:39:15Z) - Recurrent Chunking Mechanisms for Long-Text Machine Reading
Comprehension [59.80926970481975]
We study machine reading comprehension (MRC) on long texts.
A model takes as inputs a lengthy document and a question and then extracts a text span from the document as an answer.
We propose to let a model learn to chunk in a more flexible way via reinforcement learning.
arXiv Detail & Related papers (2020-05-16T18:08:58Z) - Document-Level Event Role Filler Extraction using Multi-Granularity
Contextualized Encoding [40.13163091122463]
Event extraction is a difficult task since it requires a view of a larger context to determine which spans of text correspond to event role fillers.
We first investigate how end-to-end neural sequence models perform on document-level role filler extraction.
We show that our best system performs substantially better than prior work.
arXiv Detail & Related papers (2020-05-13T20:42:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.