Answering Complex Open-Domain Questions with Multi-Hop Dense Retrieval
- URL: http://arxiv.org/abs/2009.12756v2
- Date: Fri, 19 Feb 2021 22:15:03 GMT
- Title: Answering Complex Open-Domain Questions with Multi-Hop Dense Retrieval
- Authors: Wenhan Xiong, Xiang Lorraine Li, Srini Iyer, Jingfei Du, Patrick
Lewis, William Yang Wang, Yashar Mehdad, Wen-tau Yih, Sebastian Riedel, Douwe
Kiela, Barlas O\u{g}uz
- Abstract summary: We propose a simple and efficient multi-hop dense retrieval approach for answering complex open-domain questions.
Our method does not require access to any corpus-specific information, such as inter-document hyperlinks or human-annotated entity markers.
Our system also yields a much better efficiency-accuracy trade-off, matching the best published accuracy on HotpotQA while being 10 times faster at inference time.
- Score: 117.07047313964773
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose a simple and efficient multi-hop dense retrieval approach for
answering complex open-domain questions, which achieves state-of-the-art
performance on two multi-hop datasets, HotpotQA and multi-evidence FEVER.
Contrary to previous work, our method does not require access to any
corpus-specific information, such as inter-document hyperlinks or
human-annotated entity markers, and can be applied to any unstructured text
corpus. Our system also yields a much better efficiency-accuracy trade-off,
matching the best published accuracy on HotpotQA while being 10 times faster at
inference time.
Related papers
- Adaptive Information Seeking for Open-Domain Question Answering [61.39330982757494]
We propose a novel adaptive information-seeking strategy for open-domain question answering, namely AISO.
According to the learned policy, AISO could adaptively select a proper retrieval action to seek the missing evidence at each step.
AISO outperforms all baseline methods with predefined strategies in terms of both retrieval and answer evaluations.
arXiv Detail & Related papers (2021-09-14T15:08:13Z) - Memory Augmented Sequential Paragraph Retrieval for Multi-hop Question
Answering [32.69969157825044]
We propose a new architecture that models paragraphs as sequential data and considers multi-hop information retrieval as a kind of sequence labeling task.
We evaluate our method on both full wiki and distractor subtask of HotpotQA, a public textual multi-hop QA dataset.
arXiv Detail & Related papers (2021-02-07T08:15:51Z) - Open Question Answering over Tables and Text [55.8412170633547]
In open question answering (QA), the answer to a question is produced by retrieving and then analyzing documents that might contain answers to the question.
Most open QA systems have considered only retrieving information from unstructured text.
We present a new large-scale dataset Open Table-and-Text Question Answering (OTT-QA) to evaluate performance on this task.
arXiv Detail & Related papers (2020-10-20T16:48:14Z) - Tradeoffs in Sentence Selection Techniques for Open-Domain Question
Answering [54.541952928070344]
We describe two groups of models for sentence selection: QA-based approaches, which run a full-fledged QA system to identify answer candidates, and retrieval-based models, which find parts of each passage specifically related to each question.
We show that very lightweight QA models can do well at this task, but retrieval-based models are faster still.
arXiv Detail & Related papers (2020-09-18T23:39:15Z) - Answering Any-hop Open-domain Questions with Iterative Document
Reranking [62.76025579681472]
We propose a unified QA framework to answer any-hop open-domain questions.
Our method consistently achieves performance comparable to or better than the state-of-the-art on both single-hop and multi-hop open-domain QA datasets.
arXiv Detail & Related papers (2020-09-16T04:31:38Z) - Dense Passage Retrieval for Open-Domain Question Answering [49.028342823838486]
We show that retrieval can be practically implemented using dense representations alone.
Our dense retriever outperforms a strong Lucene-BM25 system largely by 9%-19% absolute in terms of top-20 passage retrieval accuracy.
arXiv Detail & Related papers (2020-04-10T04:53:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.