EfficientQA : a RoBERTa Based Phrase-Indexed Question-Answering System
- URL: http://arxiv.org/abs/2101.02157v2
- Date: Sat, 30 Jan 2021 23:40:38 GMT
- Title: EfficientQA : a RoBERTa Based Phrase-Indexed Question-Answering System
- Authors: Sofian Chaybouti, Achraf Saghe, Aymen Shabou
- Abstract summary: In this paper, we explore the possibility to transfer the natural language understanding of language models into dense vectors representing questions and answer candidates.
Our model achieves state-of-the-art results in Phrase-Indexed Question Answering (PIQA) beating the previous state-of-art by 1.3 points in exact-match and 1.4 points in f1-score.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: State-of-the-art extractive question answering models achieve superhuman
performances on the SQuAD benchmark. Yet, they are unreasonably heavy and need
expensive GPU computing to answer questions in a reasonable time. Thus, they
cannot be used for real-world queries on hundreds of thousands of documents in
the open-domain question answering paradigm. In this paper, we explore the
possibility to transfer the natural language understanding of language models
into dense vectors representing questions and answer candidates, in order to
make the task of question-answering compatible with a simple nearest neighbor
search task. This new model, that we call EfficientQA, takes advantage from the
pair of sequences kind of input of BERT-based models to build meaningful dense
representations of candidate answers. These latter are extracted from the
context in a question-agnostic fashion. Our model achieves state-of-the-art
results in Phrase-Indexed Question Answering (PIQA) beating the previous
state-of-art by 1.3 points in exact-match and 1.4 points in f1-score. These
results show that dense vectors are able to embed very rich semantic
representations of sequences, although these ones were built from language
models not originally trained for the use-case. Thus, in order to build more
resource efficient NLP systems in the future, training language models that are
better adapted to build dense representations of phrases is one of the
possibilities.
Related papers
- HPE:Answering Complex Questions over Text by Hybrid Question Parsing and
Execution [92.69684305578957]
We propose a framework of question parsing and execution on textual QA.
The proposed framework can be viewed as a top-down question parsing followed by a bottom-up answer backtracking.
Our experiments on MuSiQue, 2WikiQA, HotpotQA, and NQ show that the proposed parsing and hybrid execution framework outperforms existing approaches in supervised, few-shot, and zero-shot settings.
arXiv Detail & Related papers (2023-05-12T22:37:06Z) - UniKGQA: Unified Retrieval and Reasoning for Solving Multi-hop Question
Answering Over Knowledge Graph [89.98762327725112]
Multi-hop Question Answering over Knowledge Graph(KGQA) aims to find the answer entities that are multiple hops away from the topic entities mentioned in a natural language question.
We propose UniKGQA, a novel approach for multi-hop KGQA task, by unifying retrieval and reasoning in both model architecture and parameter learning.
arXiv Detail & Related papers (2022-12-02T04:08:09Z) - Learn to Explain: Multimodal Reasoning via Thought Chains for Science
Question Answering [124.16250115608604]
We present Science Question Answering (SQA), a new benchmark that consists of 21k multimodal multiple choice questions with a diverse set of science topics and annotations of their answers with corresponding lectures and explanations.
We show that SQA improves the question answering performance by 1.20% in few-shot GPT-3 and 3.99% in fine-tuned UnifiedQA.
Our analysis further shows that language models, similar to humans, benefit from explanations to learn from fewer data and achieve the same performance with just 40% of the data.
arXiv Detail & Related papers (2022-09-20T07:04:24Z) - ListReader: Extracting List-form Answers for Opinion Questions [18.50111430378249]
ListReader is a neural ex-tractive QA model for list-form answer.
In addition to learning the alignment between the question and content, we introduce a heterogeneous graph neural network.
Our model adopts a co-extraction setting that can extract either span- or sentence-level answers.
arXiv Detail & Related papers (2021-10-22T10:33:08Z) - VANiLLa : Verbalized Answers in Natural Language at Large Scale [2.9098477555578333]
This dataset consists of over 100k simple questions adapted from the CSQA and SimpleQuestionsWikidata datasets.
The answer sentences in this dataset are syntactically and semantically closer to the question than to the triple fact.
arXiv Detail & Related papers (2021-05-24T16:57:54Z) - Learning Dense Representations of Phrases at Scale [22.792942611601347]
We show for the first time that we can learn dense phrase representations alone that achieve much stronger performance in open-domain QA.
Our model DensePhrases improves previous phrase retrieval models by 15%-25% absolute accuracy.
Our model is easy to parallelize due to pure dense representations and processes more than 10 questions per second on CPUs.
arXiv Detail & Related papers (2020-12-23T12:28:17Z) - Infusing Finetuning with Semantic Dependencies [62.37697048781823]
We show that, unlike syntax, semantics is not brought to the surface by today's pretrained models.
We then use convolutional graph encoders to explicitly incorporate semantic parses into task-specific finetuning.
arXiv Detail & Related papers (2020-12-10T01:27:24Z) - SPARTA: Efficient Open-Domain Question Answering via Sparse Transformer
Matching Retrieval [24.77260903221371]
We introduce SPARTA, a novel neural retrieval method that shows great promise in performance, generalization, and interpretability for open-domain question answering.
SPARTA learns a sparse representation that can be efficiently implemented as an Inverted Index.
We validated our approaches on 4 open-domain question answering (OpenQA) tasks and 11 retrieval question answering (ReQA) tasks.
arXiv Detail & Related papers (2020-09-28T02:11:02Z) - Text Modular Networks: Learning to Decompose Tasks in the Language of
Existing Models [61.480085460269514]
We propose a framework for building interpretable systems that learn to solve complex tasks by decomposing them into simpler ones solvable by existing models.
We use this framework to build ModularQA, a system that can answer multi-hop reasoning questions by decomposing them into sub-questions answerable by a neural factoid single-span QA model and a symbolic calculator.
arXiv Detail & Related papers (2020-09-01T23:45:42Z) - Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks [133.93803565077337]
retrieval-augmented generation models combine pre-trained parametric and non-parametric memory for language generation.
We show that RAG models generate more specific, diverse and factual language than a state-of-the-art parametric-only seq2seq baseline.
arXiv Detail & Related papers (2020-05-22T21:34:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.