Personalized Query Rewriting in Conversational AI Agents
- URL: http://arxiv.org/abs/2011.04748v1
- Date: Mon, 9 Nov 2020 20:45:39 GMT
- Title: Personalized Query Rewriting in Conversational AI Agents
- Authors: Alireza Roshan-Ghias, Clint Solomon Mathialagan, Pragaash Ponnusamy,
Lambert Mathias, Chenlei Guo
- Abstract summary: We propose a query rewriting approach by leveraging users' historically successful interactions as a form of memory.
We present a neural retrieval model and a pointer-generator network with hierarchical attention and show that they perform significantly better at the query rewriting task with the aforementioned user memories than without.
- Score: 7.086654234990377
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Spoken language understanding (SLU) systems in conversational AI agents often
experience errors in the form of misrecognitions by automatic speech
recognition (ASR) or semantic gaps in natural language understanding (NLU).
These errors easily translate to user frustrations, particularly so in
recurrent events e.g. regularly toggling an appliance, calling a frequent
contact, etc. In this work, we propose a query rewriting approach by leveraging
users' historically successful interactions as a form of memory. We present a
neural retrieval model and a pointer-generator network with hierarchical
attention and show that they perform significantly better at the query
rewriting task with the aforementioned user memories than without. We also
highlight how our approach with the proposed models leverages the structural
and semantic diversity in ASR's output towards recovering users' intents.
Related papers
- Do LLMs suffer from Multi-Party Hangover? A Diagnostic Approach to Addressee Recognition and Response Selection in Conversations [11.566214724241798]
We propose a methodological pipeline to investigate model performance across specific structural attributes of conversations.
We focus on Response Selection and Addressee Recognition tasks, to diagnose model weaknesses.
Results show that response selection relies more on the textual content of conversations, while addressee recognition requires capturing their structural dimension.
arXiv Detail & Related papers (2024-09-27T10:07:33Z) - Towards interfacing large language models with ASR systems using confidence measures and prompting [54.39667883394458]
This work investigates post-hoc correction of ASR transcripts with large language models (LLMs)
To avoid introducing errors into likely accurate transcripts, we propose a range of confidence-based filtering methods.
Our results indicate that this can improve the performance of less competitive ASR systems.
arXiv Detail & Related papers (2024-07-31T08:00:41Z) - Improved Contextual Recognition In Automatic Speech Recognition Systems
By Semantic Lattice Rescoring [4.819085609772069]
We propose a novel approach for enhancing contextual recognition within ASR systems via semantic lattice processing.
Our solution consists of using Hidden Markov Models and Gaussian Mixture Models (HMM-GMM) along with Deep Neural Networks (DNN) models for better accuracy.
We demonstrate the effectiveness of our proposed framework on the LibriSpeech dataset with empirical analyses.
arXiv Detail & Related papers (2023-10-14T23:16:05Z) - PICK: Polished & Informed Candidate Scoring for Knowledge-Grounded
Dialogue Systems [59.1250765143521]
Current knowledge-grounded dialogue systems often fail to align the generated responses with human-preferred qualities.
We propose Polished & Informed Candidate Scoring (PICK), a generation re-scoring framework.
We demonstrate the effectiveness of PICK in generating responses that are more faithful while keeping them relevant to the dialogue history.
arXiv Detail & Related papers (2023-09-19T08:27:09Z) - Graph Meets LLM: A Novel Approach to Collaborative Filtering for Robust
Conversational Understanding [20.934922599119865]
Defective queries can arise from user ambiguities, mistakes, or errors in automatic speech recognition (ASR) and natural language understanding (NLU)
Personalized query rewriting is an approach that focuses on reducing defects in queries by taking into account the user's individual behavior and preferences.
This paper presents our "Collaborative Query Rewriting" approach, which specifically addresses the task of rewriting new user interactions that have not been previously observed in the user's history.
arXiv Detail & Related papers (2023-05-23T18:15:29Z) - On the Impact of Speech Recognition Errors in Passage Retrieval for
Spoken Question Answering [13.013751306590303]
We study the robustness of lexical and dense retrievers against questions with synthetic ASR noise.
We create a new dataset with questions voiced by human users and use their transcriptions to show that the retrieval performance can further degrade when dealing with natural ASR noise instead of synthetic ASR noise.
arXiv Detail & Related papers (2022-09-26T18:29:36Z) - Decoupling Knowledge from Memorization: Retrieval-augmented Prompt
Learning [113.58691755215663]
We develop RetroPrompt to help a model strike a balance between generalization and memorization.
In contrast with vanilla prompt learning, RetroPrompt constructs an open-book knowledge-store from training instances.
Extensive experiments demonstrate that RetroPrompt can obtain better performance in both few-shot and zero-shot settings.
arXiv Detail & Related papers (2022-05-29T16:07:30Z) - User-Initiated Repetition-Based Recovery in Multi-Utterance Dialogue
Systems [3.20350998499235]
We present a system that allows a user to correct speech recognition errors in a virtual assistant by repeating misunderstood words.
When a user repeats part of the phrase the system rewrites the original query to incorporate the correction.
We show that rewriting the original query is an effective way to handle repetition-based recovery.
arXiv Detail & Related papers (2021-08-02T23:32:13Z) - Contextualized Attention-based Knowledge Transfer for Spoken
Conversational Question Answering [63.72278693825945]
Spoken conversational question answering (SCQA) requires machines to model complex dialogue flow.
We propose CADNet, a novel contextualized attention-based distillation approach.
We conduct extensive experiments on the Spoken-CoQA dataset and demonstrate that our approach achieves remarkable performance.
arXiv Detail & Related papers (2020-10-21T15:17:18Z) - Improving Readability for Automatic Speech Recognition Transcription [50.86019112545596]
We propose a novel NLP task called ASR post-processing for readability (APR)
APR aims to transform the noisy ASR output into a readable text for humans and downstream tasks while maintaining the semantic meaning of the speaker.
We compare fine-tuned models based on several open-sourced and adapted pre-trained models with the traditional pipeline method.
arXiv Detail & Related papers (2020-04-09T09:26:42Z) - Joint Contextual Modeling for ASR Correction and Language Understanding [60.230013453699975]
We propose multi-task neural approaches to perform contextual language correction on ASR outputs jointly with language understanding (LU)
We show that the error rates of off the shelf ASR and following LU systems can be reduced significantly by 14% relative with joint models trained using small amounts of in-domain data.
arXiv Detail & Related papers (2020-01-28T22:09:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.