Context Generation Improves Open Domain Question Answering
- URL: http://arxiv.org/abs/2210.06349v2
- Date: Thu, 27 Apr 2023 05:55:28 GMT
- Title: Context Generation Improves Open Domain Question Answering
- Authors: Dan Su, Mostofa Patwary, Shrimai Prabhumoye, Peng Xu, Ryan Prenger,
Mohammad Shoeybi, Pascale Fung, Anima Anandkumar, Bryan Catanzaro
- Abstract summary: We propose a two-stage, closed-book QA framework which employs a coarse-to-fine approach to extract relevant knowledge and answer a question.
Our method is able to better exploit the stored knowledge in pretrained LMs without adding extra learnable parameters or needing finetuning.
- Score: 102.34183939011352
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Closed-book question answering (QA) requires a model to directly answer an
open-domain question without access to any external knowledge. Prior work on
closed-book QA either directly finetunes or prompts a pretrained language model
(LM) to leverage the stored knowledge. However, they do not fully exploit the
parameterized knowledge. To address this issue, we propose a two-stage,
closed-book QA framework which employs a coarse-to-fine approach to extract
relevant knowledge and answer a question. Our approach first generates a
related context for a given question by prompting a pretrained LM. We then
prompt the same LM for answer prediction using the generated context and the
question. Additionally, to eliminate failure caused by context uncertainty, we
marginalize over generated contexts. Experimental results on three QA
benchmarks show that our method significantly outperforms previous closed-book
QA methods (e.g. exact matching 68.6% vs. 55.3%), and is on par with open-book
methods that exploit external knowledge sources (e.g. 68.6% vs. 68.0%). Our
method is able to better exploit the stored knowledge in pretrained LMs without
adding extra learnable parameters or needing finetuning, and paves the way for
hybrid models that integrate pretrained LMs with external knowledge.
Related papers
- Merging Generated and Retrieved Knowledge for Open-Domain QA [72.42262579925911]
COMBO is a compatibility-Oriented knowledge Merging for Better Open-domain QA framework.
We show that COMBO outperforms competitive baselines on three out of four tested open-domain QA benchmarks.
arXiv Detail & Related papers (2023-10-22T19:37:06Z) - A Simple Baseline for Knowledge-Based Visual Question Answering [78.00758742784532]
This paper is on the problem of Knowledge-Based Visual Question Answering (KB-VQA)
Our main contribution in this paper is to propose a much simpler and readily reproducible pipeline.
Contrary to recent approaches, our method is training-free, does not require access to external databases or APIs, and achieves state-of-the-art accuracy on the OK-VQA and A-OK-VQA datasets.
arXiv Detail & Related papers (2023-10-20T15:08:17Z) - Open-Set Knowledge-Based Visual Question Answering with Inference Paths [79.55742631375063]
The purpose of Knowledge-Based Visual Question Answering (KB-VQA) is to provide a correct answer to the question with the aid of external knowledge bases.
We propose a new retriever-ranker paradigm of KB-VQA, Graph pATH rankER (GATHER for brevity)
Specifically, it contains graph constructing, pruning, and path-level ranking, which not only retrieves accurate answers but also provides inference paths that explain the reasoning process.
arXiv Detail & Related papers (2023-10-12T09:12:50Z) - When to Read Documents or QA History: On Unified and Selective
Open-domain QA [22.941325275188376]
This paper studies the problem of open-domain question answering, with the aim of answering a diverse range of questions leveraging knowledge resources.
Two types of sources, QA-pair and document corpora, have been actively leveraged with the following complementary strength.
A natural follow-up is thus leveraging both models, while a naive pipelining or integration approaches have failed to bring additional gains over either model alone.
arXiv Detail & Related papers (2023-06-07T06:03:39Z) - Knowledge-Augmented Language Model Prompting for Zero-Shot Knowledge
Graph Question Answering [7.888547093390469]
Large Language Models (LLMs) are capable of performing zero-shot closed-book question answering tasks.
We propose to augment the knowledge directly in the input of LLMs.
Our framework, Knowledge-Augmented language model PromptING (KAPING), requires no model training, thus completely zero-shot.
arXiv Detail & Related papers (2023-06-07T04:15:21Z) - Self-Prompting Large Language Models for Zero-Shot Open-Domain QA [67.08732962244301]
Open-Domain Question Answering (ODQA) aims to answer questions without explicitly providing background documents.
This task becomes notably challenging in a zero-shot setting where no data is available to train tailored retrieval-reader models.
We propose a Self-Prompting framework to explicitly utilize the massive knowledge encoded in the parameters of Large Language Models.
arXiv Detail & Related papers (2022-12-16T18:23:43Z) - Multifaceted Improvements for Conversational Open-Domain Question
Answering [54.913313912927045]
We propose a framework with Multifaceted Improvements for Conversational open-domain Question Answering (MICQA)
Firstly, the proposed KL-divergence based regularization is able to lead to a better question understanding for retrieval and answer reading.
Second, the added post-ranker module can push more relevant passages to the top placements and be selected for reader with a two-aspect constrains.
Third, the well designed curriculum learning strategy effectively narrows the gap between the golden passage settings of training and inference, and encourages the reader to find true answer without the golden passage assistance.
arXiv Detail & Related papers (2022-04-01T07:54:27Z) - Calculating Question Similarity is Enough:A New Method for KBQA Tasks [8.056701645706404]
This paper proposes a Corpus Generation - Retrieve Method (CGRM) with Pre-training Language Model (PLM) and Knowledge Graph (KG)
Firstly, based on the mT5 model, we designed two new pre-training tasks: knowledge masked language modeling and question generation based on the paragraph.
Secondly, after preprocessing triples of knowledge graph with a series of rules, the kT5 model generates natural language QA pairs based on processed triples.
arXiv Detail & Related papers (2021-11-15T10:31:46Z) - Can Generative Pre-trained Language Models Serve as Knowledge Bases for
Closed-book QA? [10.301990059068377]
We construct a new dataset of closed-book QA using SQuAD, and investigate the performance of BART.
Experiments show that it is challenging for BART to remember training facts in high precision, and also challenging to answer closed-book questions even if relevant knowledge is retained.
arXiv Detail & Related papers (2021-06-03T03:04:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.