Learning to Simulate Natural Language Feedback for Interactive Semantic
Parsing
- URL: http://arxiv.org/abs/2305.08195v2
- Date: Sun, 4 Jun 2023 21:05:26 GMT
- Title: Learning to Simulate Natural Language Feedback for Interactive Semantic
Parsing
- Authors: Hao Yan, Saurabh Srivastava, Yintao Tai, Sida I. Wang, Wen-tau Yih,
Ziyu Yao
- Abstract summary: We propose a new task simulating NL feedback for interactive semantic parsing.
We accompany the task with a novel feedback evaluator.
Our feedback simulator can help achieve comparable error correction performance as trained using the costly, full set of human annotations.
- Score: 30.609805601567178
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Interactive semantic parsing based on natural language (NL) feedback, where
users provide feedback to correct the parser mistakes, has emerged as a more
practical scenario than the traditional one-shot semantic parsing. However,
prior work has heavily relied on human-annotated feedback data to train the
interactive semantic parser, which is prohibitively expensive and not scalable.
In this work, we propose a new task of simulating NL feedback for interactive
semantic parsing. We accompany the task with a novel feedback evaluator. The
evaluator is specifically designed to assess the quality of the simulated
feedback, based on which we decide the best feedback simulator from our
proposed variants. On a text-to-SQL dataset, we show that our feedback
simulator can generate high-quality NL feedback to boost the error correction
ability of a specific parser. In low-data settings, our feedback simulator can
help achieve comparable error correction performance as trained using the
costly, full set of human annotations.
Related papers
- Beyond Thumbs Up/Down: Untangling Challenges of Fine-Grained Feedback for Text-to-Image Generation [67.88747330066049]
Fine-grained feedback captures nuanced distinctions in image quality and prompt-alignment.
We show that demonstrating its superiority to coarse-grained feedback is not automatic.
We identify key challenges in eliciting and utilizing fine-grained feedback.
arXiv Detail & Related papers (2024-06-24T17:19:34Z) - RLVF: Learning from Verbal Feedback without Overgeneralization [94.19501420241188]
We study the problem of incorporating verbal feedback without such overgeneralization.
We develop a new method Contextualized Critiques with Constrained Preference Optimization (C3PO)
Our approach effectively applies verbal feedback to relevant scenarios while preserving existing behaviors for other contexts.
arXiv Detail & Related papers (2024-02-16T18:50:24Z) - Constructive Large Language Models Alignment with Diverse Feedback [76.9578950893839]
We introduce Constructive and Diverse Feedback (CDF) as a novel method to enhance large language models alignment.
We exploit critique feedback for easy problems, refinement feedback for medium problems, and preference feedback for hard problems.
By training our model with this diversified feedback, we achieve enhanced alignment performance while using less training data.
arXiv Detail & Related papers (2023-10-10T09:20:14Z) - UltraFeedback: Boosting Language Models with Scaled AI Feedback [99.4633351133207]
We present textscUltraFeedback, a large-scale, high-quality, and diversified AI feedback dataset.
Our work validates the effectiveness of scaled AI feedback data in constructing strong open-source chat language models.
arXiv Detail & Related papers (2023-10-02T17:40:01Z) - System-Level Natural Language Feedback [83.24259100437965]
We show how to use feedback to formalize system-level design decisions in a human-in-the-loop-process.
We conduct two case studies of this approach for improving search query and dialog response generation.
We show the combination of system-level and instance-level feedback brings further gains.
arXiv Detail & Related papers (2023-06-23T16:21:40Z) - Training Language Models with Language Feedback at Scale [50.70091340506957]
We introduce learning from Language Feedback (ILF), a new approach that utilizes more informative language feedback.
ILF consists of three steps that are applied iteratively: first, conditioning the language model on the input, an initial LM output, and feedback to generate refinements.
We show theoretically that ILF can be viewed as Bayesian Inference, similar to Reinforcement Learning from human feedback.
arXiv Detail & Related papers (2023-03-28T17:04:15Z) - NL-EDIT: Correcting semantic parse errors through natural language
interaction [28.333860779302306]
We present NL-EDIT, a model for interpreting natural language feedback in the interaction context.
We show that NL-EDIT can boost the accuracy of existing text-to-allies by up to 20% with only one turn of correction.
arXiv Detail & Related papers (2021-03-26T15:45:46Z) - Speak to your Parser: Interactive Text-to-SQL with Natural Language
Feedback [39.45695779589969]
We study the task of semantic parse correction with natural language feedback.
In this paper, we investigate a more interactive scenario where humans can further interact with the system.
arXiv Detail & Related papers (2020-05-05T23:58:09Z) - An Imitation Game for Learning Semantic Parsers from User Interaction [43.66945504686796]
We suggest an alternative, human-in-the-loop methodology for learning semantic annotations directly from users.
A semantic should be introspective and prompt for user demonstration when uncertain.
In doing so it also gets to imitate the user behavior and continue improving itself autonomously.
arXiv Detail & Related papers (2020-05-02T03:30:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.