A Coarse to Fine Question Answering System based on Reinforcement
Learning
- URL: http://arxiv.org/abs/2106.00257v1
- Date: Tue, 1 Jun 2021 06:41:48 GMT
- Title: A Coarse to Fine Question Answering System based on Reinforcement
Learning
- Authors: Yu Wang, Hongxia Jin
- Abstract summary: The system is designed using an actor-critic based deep reinforcement learning model to achieve multi-step question answering.
We test our model on four QA datasets, WIKEREADING, WIKIREADING LONG, CNN and SQuAD, and demonstrate 1.3$%$-1.7$%$ accuracy improvements with 1.5x-3.4x training speed-ups.
- Score: 48.80863342506432
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we present a coarse to fine question answering (CFQA) system
based on reinforcement learning which can efficiently processes documents with
different lengths by choosing appropriate actions. The system is designed using
an actor-critic based deep reinforcement learning model to achieve multi-step
question answering. Compared to previous QA models targeting on datasets mainly
containing either short or long documents, our multi-step coarse to fine model
takes the merits from multiple system modules, which can handle both short and
long documents. The system hence obtains a much better accuracy and faster
trainings speed compared to the current state-of-the-art models. We test our
model on four QA datasets, WIKEREADING, WIKIREADING LONG, CNN and SQuAD, and
demonstrate 1.3$\%$-1.7$\%$ accuracy improvements with 1.5x-3.4x training
speed-ups in comparison to the baselines using state-of-the-art models.
Related papers
- Evaluating Correctness and Faithfulness of Instruction-Following Models for Question Answering [26.34649731975005]
Retriever-augmented instruction-following models are attractive alternatives to fine-tuned approaches for question answering (QA)
While the model responses tend to be natural and fluent, the additional verbosity makes traditional QA evaluation metrics unreliable for accurately quantifying model performance.
We use both automatic and human evaluation to evaluate these models along two dimensions: 1) how well they satisfy the user's information need (correctness) and 2) whether they produce a response based on the provided knowledge (faithfulness)
arXiv Detail & Related papers (2023-07-31T17:41:00Z) - RAND: Robustness Aware Norm Decay For Quantized Seq2seq Models [14.07649230604283]
We propose low complexity changes to the quantization aware training (QAT) process to improve model accuracy.
With the improved accuracy, it opens up the possibility to exploit some of the other benefits of noise based QAT.
arXiv Detail & Related papers (2023-05-24T19:45:56Z) - Retrieval as Attention: End-to-end Learning of Retrieval and Reading
within a Single Transformer [80.50327229467993]
We show that a single model trained end-to-end can achieve both competitive retrieval and QA performance.
We show that end-to-end adaptation significantly boosts its performance on out-of-domain datasets in both supervised and unsupervised settings.
arXiv Detail & Related papers (2022-12-05T04:51:21Z) - KGI: An Integrated Framework for Knowledge Intensive Language Tasks [16.511913995069097]
In this paper, we propose a system based on an enhanced version of this approach for other knowledge intensive language tasks.
Our system achieves results comparable to the best models in the KILT leaderboards.
arXiv Detail & Related papers (2022-04-08T10:36:21Z) - 3M: Multi-loss, Multi-path and Multi-level Neural Networks for speech
recognition [31.992543274210835]
We identify and integrate several approaches to achieve further improvements for ASR tasks.
Specifically, multi-loss refers to the joint CTC/AED loss and multi-path denotes the Mixture-of-Experts(MoE) architecture.
We evaluate our proposed method on the public WenetSpeech dataset and experimental results show that the proposed method provides 12.2%-17.6% relative CER improvement.
arXiv Detail & Related papers (2022-04-07T03:10:49Z) - Joint Models for Answer Verification in Question Answering Systems [85.93456768689404]
We build a three-way multi-classifier, which decides if an answer supports, refutes, or is neutral with respect to another one.
We tested our models on WikiQA, TREC-QA, and a real-world dataset.
arXiv Detail & Related papers (2021-07-09T05:34:36Z) - Knowledge Distillation for Quality Estimation [79.51452598302934]
Quality Estimation (QE) is the task of automatically predicting Machine Translation quality in the absence of reference translations.
Recent success in QE stems from the use of multilingual pre-trained representations, where very large models lead to impressive results.
We show that this approach, in combination with data augmentation, leads to light-weight QE models that perform competitively with distilled pre-trained representations with 8x fewer parameters.
arXiv Detail & Related papers (2021-07-01T12:36:21Z) - When Liebig's Barrel Meets Facial Landmark Detection: A Practical Model [87.25037167380522]
We propose a model that is accurate, robust, efficient, generalizable, and end-to-end trainable.
In order to achieve a better accuracy, we propose two lightweight modules.
DQInit dynamically initializes the queries of decoder from the inputs, enabling the model to achieve as good accuracy as the ones with multiple decoder layers.
QAMem is designed to enhance the discriminative ability of queries on low-resolution feature maps by assigning separate memory values to each query rather than a shared one.
arXiv Detail & Related papers (2021-05-27T13:51:42Z) - SRQA: Synthetic Reader for Factoid Question Answering [21.28441702154528]
We introduce a new model called SRQA, which means Synthetic Reader for Factoid Question Answering.
This model enhances the question answering system in the multi-document scenario from three aspects.
We perform SRQA on the WebQA dataset, and experiments show that our model outperforms the state-of-the-art models.
arXiv Detail & Related papers (2020-09-02T13:16:24Z) - The Right Tool for the Job: Matching Model and Instance Complexities [62.95183777679024]
As NLP models become larger, executing a trained model requires significant computational resources incurring monetary and environmental costs.
We propose a modification to contextual representation fine-tuning which, during inference, allows for an early (and fast) "exit"
We test our proposed modification on five different datasets in two tasks: three text classification datasets and two natural language inference benchmarks.
arXiv Detail & Related papers (2020-04-16T04:28:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.