Answering Unanswered Questions through Semantic Reformulations in Spoken
QA
- URL: http://arxiv.org/abs/2305.17393v2
- Date: Sat, 3 Jun 2023 05:40:25 GMT
- Title: Answering Unanswered Questions through Semantic Reformulations in Spoken
QA
- Authors: Pedro Faustini, Zhiyu Chen, Besnik Fetahu, Oleg Rokhlenko and Shervin
Malmasi
- Abstract summary: Spoken Question Answering (QA) is a key feature of voice assistants, usually backed by multiple QA systems.
We analyze failed QA requests to identify core challenges: lexical gaps, proposition types, complex syntactic structure, and high specificity.
We propose a Semantic Question Reformulation (SURF) model offering three linguistically-grounded operations (repair, syntactic reshaping, generalization) to rewrite questions to facilitate answering.
- Score: 20.216161323866867
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Spoken Question Answering (QA) is a key feature of voice assistants, usually
backed by multiple QA systems. Users ask questions via spontaneous speech which
can contain disfluencies, errors, and informal syntax or phrasing. This is a
major challenge in QA, causing unanswered questions or irrelevant answers, and
leading to bad user experiences. We analyze failed QA requests to identify core
challenges: lexical gaps, proposition types, complex syntactic structure, and
high specificity. We propose a Semantic Question Reformulation (SURF) model
offering three linguistically-grounded operations (repair, syntactic reshaping,
generalization) to rewrite questions to facilitate answering. Offline
evaluation on 1M unanswered questions from a leading voice assistant shows that
SURF significantly improves answer rates: up to 24% of previously unanswered
questions obtain relevant answers (75%). Live deployment shows positive impact
for millions of customers with unanswered questions; explicit relevance
feedback shows high user satisfaction.
Related papers
- Which questions should I answer? Salience Prediction of Inquisitive Questions [118.097974193544]
We show that highly salient questions are empirically more likely to be answered in the same article.
We further validate our findings by showing that answering salient questions is an indicator of summarization quality in news.
arXiv Detail & Related papers (2024-04-16T21:33:05Z) - Estimating the Usefulness of Clarifying Questions and Answers for
Conversational Search [17.0363715044341]
We propose a method for processing answers to clarifying questions, moving away from previous work that simply appends answers to the original query.
Specifically, we propose a classifier for assessing usefulness of the prompted clarifying question and an answer given by the user.
Results demonstrate significant improvements over strong non-mixed-initiative baselines.
arXiv Detail & Related papers (2024-01-21T11:04:30Z) - Answering Ambiguous Questions with a Database of Questions, Answers, and
Revisions [95.92276099234344]
We present a new state-of-the-art for answering ambiguous questions that exploits a database of unambiguous questions generated from Wikipedia.
Our method improves performance by 15% on recall measures and 10% on measures which evaluate disambiguating questions from predicted outputs.
arXiv Detail & Related papers (2023-08-16T20:23:16Z) - HeySQuAD: A Spoken Question Answering Dataset [2.3881849082514153]
This study presents a new large-scale community-shared SQA dataset called HeySQuAD.
Our goal is to measure the ability of machines to accurately understand noisy spoken questions and provide reliable answers.
arXiv Detail & Related papers (2023-04-26T17:15:39Z) - Improving the Question Answering Quality using Answer Candidate
Filtering based on Natural-Language Features [117.44028458220427]
We address the problem of how the Question Answering (QA) quality of a given system can be improved.
Our main contribution is an approach capable of identifying wrong answers provided by a QA system.
In particular, our approach has shown its potential while removing in many cases the majority of incorrect answers.
arXiv Detail & Related papers (2021-12-10T11:09:44Z) - GooAQ: Open Question Answering with Diverse Answer Types [63.06454855313667]
We present GooAQ, a large-scale dataset with a variety of answer types.
This dataset contains over 5 million questions and 3 million answers collected from Google.
arXiv Detail & Related papers (2021-04-18T05:40:39Z) - NoiseQA: Challenge Set Evaluation for User-Centric Question Answering [68.67783808426292]
We show that components in the pipeline that precede an answering engine can introduce varied and considerable sources of error.
We conclude that there is substantial room for progress before QA systems can be effectively deployed.
arXiv Detail & Related papers (2021-02-16T18:35:29Z) - Review-guided Helpful Answer Identification in E-commerce [38.276241153439955]
Product-specific community question answering platforms can greatly help address the concerns of potential customers.
The user-provided answers on such platforms often vary a lot in their qualities.
Helpfulness votes from the community can indicate the overall quality of the answer, but they are often missing.
arXiv Detail & Related papers (2020-03-13T11:34:29Z) - Unsupervised Question Decomposition for Question Answering [102.56966847404287]
We propose an algorithm for One-to-N Unsupervised Sequence Sequence (ONUS) that learns to map one hard, multi-hop question to many simpler, single-hop sub-questions.
We show large QA improvements on HotpotQA over a strong baseline on the original, out-of-domain, and multi-hop dev sets.
arXiv Detail & Related papers (2020-02-22T19:40:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.