Knowledge-Aided Open-Domain Question Answering
- URL: http://arxiv.org/abs/2006.05244v1
- Date: Tue, 9 Jun 2020 13:28:57 GMT
- Title: Knowledge-Aided Open-Domain Question Answering
- Authors: Mantong Zhou, Zhouxing Shi, Minlie Huang, Xiaoyan Zhu
- Abstract summary: We propose a knowledge-aided open-domain QA (KAQA) method which targets at improving relevant document retrieval and answer reranking.
During document retrieval, a candidate document is scored by considering its relationship to the question and other documents.
During answer reranking, a candidate answer is reranked using not only its own context but also the clues from other documents.
- Score: 58.712857964048446
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Open-domain question answering (QA) aims to find the answer to a question
from a large collection of documents.Though many models for single-document
machine comprehension have achieved strong performance, there is still much
room for improving open-domain QA systems since document retrieval and answer
reranking are still unsatisfactory. Golden documents that contain the correct
answers may not be correctly scored by the retrieval component, and the correct
answers that have been extracted may be wrongly ranked after other candidate
answers by the reranking component. One of the reasons is derived from the
independent principle in which each candidate document (or answer) is scored
independently without considering its relationship to other documents (or
answers). In this work, we propose a knowledge-aided open-domain QA (KAQA)
method which targets at improving relevant document retrieval and candidate
answer reranking by considering the relationship between a question and the
documents (termed as question-document graph), and the relationship between
candidate documents (termed as document-document graph). The graphs are built
using knowledge triples from external knowledge resources. During document
retrieval, a candidate document is scored by considering its relationship to
the question and other documents. During answer reranking, a candidate answer
is reranked using not only its own context but also the clues from other
documents. The experimental results show that our proposed method improves
document retrieval and answer reranking, and thereby enhances the overall
performance of open-domain question answering.
Related papers
- Entity Retrieval for Answering Entity-Centric Questions [4.327763441385372]
We propose a novel retrieval method which rather than relying on question-document similarity, depends on the salient entities within the question to identify the retrieval documents.
Our findings reveal that our method not only leads to more accurate answers to entity-centric questions but also operates more efficiently.
arXiv Detail & Related papers (2024-08-05T19:23:20Z) - CAPSTONE: Curriculum Sampling for Dense Retrieval with Document
Expansion [68.19934563919192]
We propose a curriculum sampling strategy that utilizes pseudo queries during training and progressively enhances the relevance between the generated query and the real query.
Experimental results on both in-domain and out-of-domain datasets demonstrate that our approach outperforms previous dense retrieval models.
arXiv Detail & Related papers (2022-12-18T15:57:46Z) - Generate rather than Retrieve: Large Language Models are Strong Context
Generators [74.87021992611672]
We present a novel perspective for solving knowledge-intensive tasks by replacing document retrievers with large language model generators.
We call our method generate-then-read (GenRead), which first prompts a large language model to generate contextutal documents based on a given question, and then reads the generated documents to produce the final answer.
arXiv Detail & Related papers (2022-09-21T01:30:59Z) - Asking questions on handwritten document collections [35.85762649504866]
This work addresses the problem of Question Answering (QA) on handwritten document collections.
Unlike typical QA and Visual Question Answering (VQA) formulations where the answer is a short text, we aim to locate a document snippet where the answer lies.
We argue that the recognition-free approach is suitable for handwritten documents and historical collections where robust text recognition is often difficult.
arXiv Detail & Related papers (2021-10-02T02:40:40Z) - Adaptive Information Seeking for Open-Domain Question Answering [61.39330982757494]
We propose a novel adaptive information-seeking strategy for open-domain question answering, namely AISO.
According to the learned policy, AISO could adaptively select a proper retrieval action to seek the missing evidence at each step.
AISO outperforms all baseline methods with predefined strategies in terms of both retrieval and answer evaluations.
arXiv Detail & Related papers (2021-09-14T15:08:13Z) - End-to-End Multihop Retrieval for Compositional Question Answering over
Long Documents [93.55268936974971]
We propose a multi-hop retrieval method, DocHopper, to answer compositional questions over long documents.
At each step, DocHopper retrieves a paragraph or sentence embedding from the document, mixes the retrieved result with the query, and updates the query for the next step.
We demonstrate that utilizing document structure in this was can largely improve question-answering and retrieval performance on long documents.
arXiv Detail & Related papers (2021-06-01T03:13:35Z) - Answering Any-hop Open-domain Questions with Iterative Document
Reranking [62.76025579681472]
We propose a unified QA framework to answer any-hop open-domain questions.
Our method consistently achieves performance comparable to or better than the state-of-the-art on both single-hop and multi-hop open-domain QA datasets.
arXiv Detail & Related papers (2020-09-16T04:31:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.