A Survey for Efficient Open Domain Question Answering
- URL: http://arxiv.org/abs/2211.07886v1
- Date: Tue, 15 Nov 2022 04:18:53 GMT
- Title: A Survey for Efficient Open Domain Question Answering
- Authors: Qin Zhang, Shangsi Chen, Dongkuan Xu, Qingqing Cao, Xiaojun Chen,
Trevor Cohn, Meng Fang
- Abstract summary: Open domain question answering (ODQA) is a longstanding task aimed at answering factual questions from a large knowledge corpus without any explicit evidence in natural language processing (NLP)
- Score: 51.67110249787223
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Open domain question answering (ODQA) is a longstanding task aimed at
answering factual questions from a large knowledge corpus without any explicit
evidence in natural language processing (NLP). Recent works have predominantly
focused on improving the answering accuracy and achieved promising progress.
However, higher accuracy often comes with more memory consumption and inference
latency, which might not necessarily be efficient enough for direct deployment
in the real world. Thus, a trade-off between accuracy, memory consumption and
processing speed is pursued. In this paper, we provide a survey of recent
advances in the efficiency of ODQA models. We walk through the ODQA models and
conclude the core techniques on efficiency. Quantitative analysis on memory
cost, processing speed, accuracy and overall comparison are given. We hope that
this work would keep interested scholars informed of the advances and open
challenges in ODQA efficiency research, and thus contribute to the further
development of ODQA efficiency.
Related papers
- EfficientEQA: An Efficient Approach for Open Vocabulary Embodied Question Answering [21.114403949257934]
Embodied Question Answering (EQA) is an essential yet challenging task for robotic home assistants.
Recent studies have shown that large vision-language models (VLMs) can be effectively utilized for EQA, but existing works either focus on video-based question answering or rely on closed-form choice sets.
We propose a novel framework called EfficientEQA for open-vocabulary EQA, which enables efficient exploration and accurate answering.
arXiv Detail & Related papers (2024-10-26T19:48:47Z) - CuriousLLM: Elevating Multi-Document QA with Reasoning-Infused Knowledge Graph Prompting [0.0]
We improve over a novel approach called Knowledge Graph Prompting (KGP), which combines knowledge graphs with a LLM-based agent to improve reasoning and search accuracy.
We propose a reasoning-infused LLM agent to enhance this framework.
This agent mimics human curiosity to ask follow-up questions to more efficiently navigate the search.
arXiv Detail & Related papers (2024-04-13T20:43:46Z) - EEE-QA: Exploring Effective and Efficient Question-Answer
Representations [7.764629726412793]
Current approaches to question answering rely on pre-trained language models (PLMs) like RoBERTa.
This work challenges the existing question-answer encoding convention and explores finer representations.
arXiv Detail & Related papers (2024-03-04T16:21:13Z) - Strong and Efficient Baselines for Open Domain Conversational Question
Answering [2.773656427800412]
We study the State-of-the-Art (SotA) Dense Passage Retrieval (DPR) retriever and Fusion-in-Decoder (FiD) reader pipeline.
We propose and evaluate strong yet simple and efficient baselines, by introducing a fast reranking component between the retriever and the reader.
Experiments on two ODConvQA tasks, namely TopiOCQA and OR-QuAC, show that our method improves the SotA results, while reducing reader's latency by 60%.
arXiv Detail & Related papers (2023-10-23T08:48:14Z) - MinPrompt: Graph-based Minimal Prompt Data Augmentation for Few-shot Question Answering [64.6741991162092]
We present MinPrompt, a minimal data augmentation framework for open-domain question answering.
We transform the raw text into a graph structure to build connections between different factual sentences.
We then apply graph algorithms to identify the minimal set of sentences needed to cover the most information in the raw text.
We generate QA pairs based on the identified sentence subset and train the model on the selected sentences to obtain the final model.
arXiv Detail & Related papers (2023-10-08T04:44:36Z) - Building Interpretable and Reliable Open Information Retriever for New
Domains Overnight [67.03842581848299]
Information retrieval is a critical component for many down-stream tasks such as open-domain question answering (QA)
We propose an information retrieval pipeline that uses entity/event linking model and query decomposition model to focus more accurately on different information units of the query.
We show that, while being more interpretable and reliable, our proposed pipeline significantly improves passage coverages and denotation accuracies across five IR and QA benchmarks.
arXiv Detail & Related papers (2023-08-09T07:47:17Z) - Improving Text Matching in E-Commerce Search with A Rationalizable,
Intervenable and Fast Entity-Based Relevance Model [78.80174696043021]
We propose a novel model called the Entity-Based Relevance Model (EBRM)
The decomposition allows us to use a Cross-encoder QE relevance module for high accuracy.
We also show that pretraining the QE module with auto-generated QE data from user logs can further improve the overall performance.
arXiv Detail & Related papers (2023-07-01T15:44:53Z) - Efficient Few-Shot Object Detection via Knowledge Inheritance [62.36414544915032]
Few-shot object detection (FSOD) aims at learning a generic detector that can adapt to unseen tasks with scarce training samples.
We present an efficient pretrain-transfer framework (PTF) baseline with no computational increment.
We also propose an adaptive length re-scaling (ALR) strategy to alleviate the vector length inconsistency between the predicted novel weights and the pretrained base weights.
arXiv Detail & Related papers (2022-03-23T06:24:31Z) - Simple and Efficient ways to Improve REALM [24.182528719468227]
REALM (Guu et al., 2020) is an end-to-end dense retrieval system that relies on pretraining for improved downstream QA efficiency.
Dense has been shown to be effective for retrieving relevant documents for Open Domain QA, surpassing popular sparse retrieval methods.
arXiv Detail & Related papers (2021-04-18T04:32:33Z) - Harvesting and Refining Question-Answer Pairs for Unsupervised QA [95.9105154311491]
We introduce two approaches to improve unsupervised Question Answering (QA)
First, we harvest lexically and syntactically divergent questions from Wikipedia to automatically construct a corpus of question-answer pairs (named as RefQA)
Second, we take advantage of the QA model to extract more appropriate answers, which iteratively refines data over RefQA.
arXiv Detail & Related papers (2020-05-06T15:56:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.