Improving Conversational Question Answering Systems after Deployment
using Feedback-Weighted Learning
- URL: http://arxiv.org/abs/2011.00615v1
- Date: Sun, 1 Nov 2020 19:50:34 GMT
- Title: Improving Conversational Question Answering Systems after Deployment
using Feedback-Weighted Learning
- Authors: Jon Ander Campos, Kyunghyun Cho, Arantxa Otegi, Aitor Soroa, Gorka
Azkune, Eneko Agirre
- Abstract summary: We propose feedback-weighted learning based on importance sampling to improve upon an initial supervised system using binary user feedback.
Our work opens the prospect to exploit interactions with real users and improve conversational systems after deployment.
- Score: 69.42679922160684
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The interaction of conversational systems with users poses an exciting
opportunity for improving them after deployment, but little evidence has been
provided of its feasibility. In most applications, users are not able to
provide the correct answer to the system, but they are able to provide binary
(correct, incorrect) feedback. In this paper we propose feedback-weighted
learning based on importance sampling to improve upon an initial supervised
system using binary user feedback. We perform simulated experiments on document
classification (for development) and Conversational Question Answering datasets
like QuAC and DoQA, where binary user feedback is derived from gold
annotations. The results show that our method is able to improve over the
initial supervised system, getting close to a fully-supervised system that has
access to the same labeled examples in in-domain experiments (QuAC), and even
matching in out-of-domain experiments (DoQA). Our work opens the prospect to
exploit interactions with real users and improve conversational systems after
deployment.
Related papers
- Rethinking the Evaluation of Dialogue Systems: Effects of User Feedback on Crowdworkers and LLMs [57.16442740983528]
In ad-hoc retrieval, evaluation relies heavily on user actions, including implicit feedback.
The role of user feedback in annotators' assessment of turns in a conversational perception has been little studied.
We focus on how the evaluation of task-oriented dialogue systems ( TDSs) is affected by considering user feedback, explicit or implicit, as provided through the follow-up utterance of a turn being evaluated.
arXiv Detail & Related papers (2024-04-19T16:45:50Z) - System-Level Natural Language Feedback [83.24259100437965]
We show how to use feedback to formalize system-level design decisions in a human-in-the-loop-process.
We conduct two case studies of this approach for improving search query and dialog response generation.
We show the combination of system-level and instance-level feedback brings further gains.
arXiv Detail & Related papers (2023-06-23T16:21:40Z) - Continually Improving Extractive QA via Human Feedback [59.49549491725224]
We study continually improving an extractive question answering (QA) system via human user feedback.
We conduct experiments involving thousands of user interactions under diverse setups to broaden the understanding of learning from feedback over time.
arXiv Detail & Related papers (2023-05-21T14:35:32Z) - Towards Teachable Reasoning Systems [29.59387051046722]
We develop a teachable reasoning system for question-answering (QA)
Our approach is three-fold: First, generated chains of reasoning show how answers are implied by the system's own internal beliefs.
Second, users can interact with the explanations to identify erroneous model beliefs and provide corrections.
Third, we augment the model with a dynamic memory of such corrections.
arXiv Detail & Related papers (2022-04-27T17:15:07Z) - Using Interactive Feedback to Improve the Accuracy and Explainability of
Question Answering Systems Post-Deployment [20.601284299825895]
We focus on two kinds of improvements: 1) improving the QA system's performance itself, and 2) providing the model with the ability to explain the correctness or incorrectness of an answer.
We collect a retrieval-based QA dataset, FeedbackQA, which contains interactive feedback from users.
We show that feedback data not only improves the accuracy of the deployed QA system but also other stronger non-deployed systems.
arXiv Detail & Related papers (2022-04-06T18:17:09Z) - Simulating Bandit Learning from User Feedback for Extractive Question
Answering [51.97943858898579]
We study learning from user feedback for extractive question answering by simulating feedback using supervised data.
We show that systems initially trained on a small number of examples can dramatically improve given feedback from users on model-predicted answers.
arXiv Detail & Related papers (2022-03-18T17:47:58Z) - An Adversarial Learning based Multi-Step Spoken Language Understanding
System through Human-Computer Interaction [70.25183730482915]
We introduce a novel multi-step spoken language understanding system based on adversarial learning.
We demonstrate that the new system can improve parsing performance by at least $2.5%$ in terms of F1.
arXiv Detail & Related papers (2021-06-06T03:46:53Z) - Soliciting Human-in-the-Loop User Feedback for Interactive Machine
Learning Reduces User Trust and Impressions of Model Accuracy [8.11839312231511]
Mixed-initiative systems allow users to interactively provide feedback to improve system performance.
Our research investigates how the act of providing feedback can affect user understanding of an intelligent system and its accuracy.
arXiv Detail & Related papers (2020-08-28T16:46:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.