Open-Domain Question Answering with Pre-Constructed Question Spaces
- URL: http://arxiv.org/abs/2006.08337v2
- Date: Fri, 23 Oct 2020 02:10:07 GMT
- Title: Open-Domain Question Answering with Pre-Constructed Question Spaces
- Authors: Jinfeng Xiao, Lidan Wang, Franck Dernoncourt, Trung Bui, Tong Sun,
Jiawei Han
- Abstract summary: Open-domain question answering aims at solving the task of locating the answers to user-generated questions in massive collections of documents.
There are two families of solutions available: retriever-readers, and knowledge-graph-based approaches.
We propose a novel algorithm with a reader-retriever structure that differs from both families.
- Score: 70.13619499853756
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Open-domain question answering aims at solving the task of locating the
answers to user-generated questions in massive collections of documents. There
are two families of solutions available: retriever-readers, and
knowledge-graph-based approaches. A retriever-reader usually first uses
information retrieval methods like TF-IDF to locate some documents or
paragraphs that are likely to be relevant to the question, and then feeds the
retrieved text to a neural network reader to extract the answer. Alternatively,
knowledge graphs can be constructed from the corpus and be queried against to
answer user questions. We propose a novel algorithm with a reader-retriever
structure that differs from both families. Our reader-retriever first uses an
offline reader to read the corpus and generate collections of all answerable
questions associated with their answers, and then uses an online retriever to
respond to user queries by searching the pre-constructed question spaces for
answers that are most likely to be asked in the given way. We further combine
retriever-reader and reader-retriever results into one single answer by
examining the consistency between the two components. We claim that our
algorithm solves some bottlenecks in existing work, and demonstrate that it
achieves superior accuracy on real-world datasets.
Related papers
- Phrase Retrieval for Open-Domain Conversational Question Answering with
Conversational Dependency Modeling via Contrastive Learning [54.55643652781891]
Open-Domain Conversational Question Answering (ODConvQA) aims at answering questions through a multi-turn conversation.
We propose a method to directly predict answers with a phrase retrieval scheme for a sequence of words.
arXiv Detail & Related papers (2023-06-07T09:46:38Z) - Grape: Knowledge Graph Enhanced Passage Reader for Open-domain Question
Answering [36.85435188308579]
A common thread of open-domain question answering (QA) models employs a retriever-reader pipeline that first retrieves a handful of relevant passages from Wikipedia.
We propose a novel knowledge Graph enhanced passage reader, namely Grape, to improve the reader performance for open-domain QA.
arXiv Detail & Related papers (2022-10-06T14:06:04Z) - Generate rather than Retrieve: Large Language Models are Strong Context
Generators [74.87021992611672]
We present a novel perspective for solving knowledge-intensive tasks by replacing document retrievers with large language model generators.
We call our method generate-then-read (GenRead), which first prompts a large language model to generate contextutal documents based on a given question, and then reads the generated documents to produce the final answer.
arXiv Detail & Related papers (2022-09-21T01:30:59Z) - Zero-Shot Open-Book Question Answering [0.0]
This article proposes a solution for answering natural language questions from technical documents with no domain-specific labeled data (zero-shot)
We are introducing a new test dataset for open-book QA based on real customer questions on AWS technical documentation.
We were able to achieve 49% F1 and 39% exact score (EM) end-to-end with no domain-specific training.
arXiv Detail & Related papers (2021-11-22T20:38:41Z) - Tackling Multi-Answer Open-Domain Questions via a Recall-then-Verify
Framework [38.807388762378444]
Open domain questions are likely to be open-ended and ambiguous, leading to multiple valid answers.
Existing approaches typically adopt the rerank-then-read framework, where a reader reads top-ranking evidence to predict answers.
Our framework achieves new state-of-the-art results on two multi-answer datasets, and predicts significantly more gold answers than a rerank-then-read system with an oracle reranker.
arXiv Detail & Related papers (2021-10-16T10:48:10Z) - Is Retriever Merely an Approximator of Reader? [27.306407064073177]
We show that the reader and the retriever are complementary to each other even in terms of accuracy only.
We propose to distill the reader into the retriever so that the retriever absorbs the strength of the reader while keeping its own benefit.
arXiv Detail & Related papers (2020-10-21T13:40:15Z) - Open Question Answering over Tables and Text [55.8412170633547]
In open question answering (QA), the answer to a question is produced by retrieving and then analyzing documents that might contain answers to the question.
Most open QA systems have considered only retrieving information from unstructured text.
We present a new large-scale dataset Open Table-and-Text Question Answering (OTT-QA) to evaluate performance on this task.
arXiv Detail & Related papers (2020-10-20T16:48:14Z) - Answering Any-hop Open-domain Questions with Iterative Document
Reranking [62.76025579681472]
We propose a unified QA framework to answer any-hop open-domain questions.
Our method consistently achieves performance comparable to or better than the state-of-the-art on both single-hop and multi-hop open-domain QA datasets.
arXiv Detail & Related papers (2020-09-16T04:31:38Z) - Knowledge-Aided Open-Domain Question Answering [58.712857964048446]
We propose a knowledge-aided open-domain QA (KAQA) method which targets at improving relevant document retrieval and answer reranking.
During document retrieval, a candidate document is scored by considering its relationship to the question and other documents.
During answer reranking, a candidate answer is reranked using not only its own context but also the clues from other documents.
arXiv Detail & Related papers (2020-06-09T13:28:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.