RFiD: Towards Rational Fusion-in-Decoder for Open-Domain Question
Answering
- URL: http://arxiv.org/abs/2305.17041v1
- Date: Fri, 26 May 2023 15:51:25 GMT
- Title: RFiD: Towards Rational Fusion-in-Decoder for Open-Domain Question
Answering
- Authors: Cunxiang Wang, Haofei Yu, Yue Zhang
- Abstract summary: Open-Domain Question Answering (ODQA) systems necessitate a reader model capable of generating answers by simultaneously referring to multiple passages.
We introduce the Rational Fusion-in-Decoder (RFiD) model to counter this problem.
- Score: 11.62870729875824
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Open-Domain Question Answering (ODQA) systems necessitate a reader model
capable of generating answers by simultaneously referring to multiple passages.
Although representative models like Fusion-in-Decoder (FiD) have been proposed
to address this challenge, these systems can inadvertently rely on spurious
features instead of genuine causal relationships between the question and the
passages to generate answers. To counter this problem, we introduce the
Rational Fusion-in-Decoder (RFiD) model. Our model leverages the encoders of
FiD to differentiate between causal relationships and spurious features,
subsequently guiding the decoder to generate answers informed by this
discernment. Experimental results on two ODQA datasets, Natural Questions (NQ)
and TriviaQA (TQ), demonstrate that our model surpasses previous methods,
achieving improvements of up to 1.5 and 0.7 in Exact Match scores on NQ, and
exhibits an enhanced ability to identify causal relationships.
Related papers
- Multi-Granularity Guided Fusion-in-Decoder [7.87348193562399]
We propose the Multi-Granularity guided Fusion-in-Decoder (MGFiD) to discerning evidence across multiple levels of granularity.
Based on multi-task learning, MGFiD harmonizes passage re-ranking with sentence classification.
It improves decoding efficiency by reusing the results of passage re-ranking for passage pruning.
arXiv Detail & Related papers (2024-04-03T08:56:00Z) - RQ-RAG: Learning to Refine Queries for Retrieval Augmented Generation [42.82192656794179]
Large Language Models (LLMs) exhibit remarkable capabilities but are prone to generating inaccurate or hallucinatory responses.
This limitation stems from their reliance on vast pretraining datasets, making them susceptible to errors in unseen scenarios.
Retrieval-Augmented Generation (RAG) addresses this by incorporating external, relevant documents into the response generation process.
arXiv Detail & Related papers (2024-03-31T08:58:54Z) - UNK-VQA: A Dataset and a Probe into the Abstention Ability of Multi-modal Large Models [55.22048505787125]
This paper contributes a comprehensive dataset, called UNK-VQA.
We first augment the existing data via deliberate perturbations on either the image or question.
We then extensively evaluate the zero- and few-shot performance of several emerging multi-modal large models.
arXiv Detail & Related papers (2023-10-17T02:38:09Z) - Toward Unsupervised Realistic Visual Question Answering [70.67698100148414]
We study the problem of realistic VQA (RVQA), where a model has to reject unanswerable questions (UQs) and answer answerable ones (AQs)
We first point out 2 drawbacks in current RVQA research, where (1) datasets contain too many unchallenging UQs and (2) a large number of annotated UQs are required for training.
We propose a new testing dataset, RGQA, which combines AQs from an existing VQA dataset with around 29K human-annotated UQs.
This combines pseudo UQs obtained by randomly pairing images and questions, with an
arXiv Detail & Related papers (2023-03-09T06:58:29Z) - Open-domain Question Answering via Chain of Reasoning over Heterogeneous
Knowledge [82.5582220249183]
We propose a novel open-domain question answering (ODQA) framework for answering single/multi-hop questions across heterogeneous knowledge sources.
Unlike previous methods that solely rely on the retriever for gathering all evidence in isolation, our intermediary performs a chain of reasoning over the retrieved set.
Our system achieves competitive performance on two ODQA datasets, OTT-QA and NQ, against tables and passages from Wikipedia.
arXiv Detail & Related papers (2022-10-22T03:21:32Z) - Enhancing Dual-Encoders with Question and Answer Cross-Embeddings for
Answer Retrieval [29.16807969384253]
Dual-Encoders is a promising mechanism for answer retrieval in question answering (QA) systems.
We propose a framework to enhance the Dual-Encoders model with question answer cross-embeddings and a novel Geometry Alignment Mechanism (GAM)
Our framework significantly improves Dual-Encoders model and outperforms the state-of-the-art method on multiple answer retrieval datasets.
arXiv Detail & Related papers (2022-06-07T02:39:24Z) - KG-FiD: Infusing Knowledge Graph in Fusion-in-Decoder for Open-Domain
Question Answering [68.00631278030627]
We propose a novel method KG-FiD, which filters noisy passages by leveraging the structural relationship among the retrieved passages with a knowledge graph.
We show that KG-FiD can improve vanilla FiD by up to 1.5% on answer exact match score and achieve comparable performance with FiD with only 40% of computation cost.
arXiv Detail & Related papers (2021-10-08T18:39:59Z) - Answering Ambiguous Questions through Generative Evidence Fusion and
Round-Trip Prediction [46.38201136570501]
We present a model that aggregates and combines evidence from multiple passages to adaptively predict a single answer or a set of question-answer pairs for ambiguous questions.
Our model, named Refuel, achieves a new state-of-the-art performance on the AmbigQA dataset, and shows competitive performance on NQ-Open and TriviaQA.
arXiv Detail & Related papers (2020-11-26T05:48:55Z) - Leveraging Passage Retrieval with Generative Models for Open Domain
Question Answering [61.394478670089065]
Generative models for open domain question answering have proven to be competitive, without resorting to external knowledge.
We investigate how much these models can benefit from retrieving text passages, potentially containing evidence.
We observe that the performance of this method significantly improves when increasing the number of retrieved passages.
arXiv Detail & Related papers (2020-07-02T17:44:57Z) - Harvesting and Refining Question-Answer Pairs for Unsupervised QA [95.9105154311491]
We introduce two approaches to improve unsupervised Question Answering (QA)
First, we harvest lexically and syntactically divergent questions from Wikipedia to automatically construct a corpus of question-answer pairs (named as RefQA)
Second, we take advantage of the QA model to extract more appropriate answers, which iteratively refines data over RefQA.
arXiv Detail & Related papers (2020-05-06T15:56:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.