Augmenting Multi-Turn Text-to-SQL Datasets with Self-Play
- URL: http://arxiv.org/abs/2210.12096v1
- Date: Fri, 21 Oct 2022 16:40:07 GMT
- Title: Augmenting Multi-Turn Text-to-SQL Datasets with Self-Play
- Authors: Qi Liu, Zihuiwen Ye, Tao Yu, Phil Blunsom, Linfeng Song
- Abstract summary: We explore augmenting the training datasets using self-play, which leverages contextual information to synthesize new interactions.
We find that self-play improves the accuracy of a strong baseline on SParC and Co, two widely used text-to-domain datasets.
- Score: 46.07002748587857
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The task of context-dependent text-to-SQL aims to convert multi-turn user
utterances to formal SQL queries. This is a challenging task due to both the
scarcity of training data from which to learn complex contextual dependencies
and to generalize to unseen databases. In this paper we explore augmenting the
training datasets using self-play, which leverages contextual information to
synthesize new interactions to adapt the model to new databases. We first
design a SQL-to-text model conditioned on a sampled goal query, which
represents a user's intent, that then converses with a text-to-SQL semantic
parser to generate new interactions. We then filter the synthesized
interactions and retrain the models with the augmented data. We find that
self-play improves the accuracy of a strong baseline on SParC and CoSQL, two
widely used cross-domain text-to-SQL datasets. Our analysis shows that
self-play simulates various conversational thematic relations, enhances
cross-domain generalization and improves beam-search.
Related papers
- SQL-GEN: Bridging the Dialect Gap for Text-to-SQL Via Synthetic Data And Model Merging [30.306023265985658]
We introduce a framework for generating high-quality synthetic training data for any dialect.
We propose a novel Mixture-of-Experts (MoE) that leverages the shared knowledge across dialects.
arXiv Detail & Related papers (2024-08-22T20:50:48Z) - SQL-PaLM: Improved Large Language Model Adaptation for Text-to-SQL (extended) [53.95151604061761]
This paper introduces the framework for enhancing Text-to- filtering using large language models (LLMs)
With few-shot prompting, we explore the effectiveness of consistency decoding with execution-based error analyses.
With instruction fine-tuning, we delve deep in understanding the critical paradigms that influence the performance of tuned LLMs.
arXiv Detail & Related papers (2023-05-26T21:39:05Z) - Importance of Synthesizing High-quality Data for Text-to-SQL Parsing [71.02856634369174]
State-of-the-art text-to-weighted algorithms did not further improve on popular benchmarks when trained with augmented synthetic data.
We propose a novel framework that incorporates key relationships from schema, imposes strong typing, and schema-weighted column sampling.
arXiv Detail & Related papers (2022-12-17T02:53:21Z) - Diverse Parallel Data Synthesis for Cross-Database Adaptation of
Text-to-SQL Parsers [21.272952382662215]
Adapting to new databases is a challenging problem due to the lack of natural language queries in the new schemas.
We present ReFill, a framework for adapting a Text-to-edit to a target schema.
arXiv Detail & Related papers (2022-10-29T14:30:53Z) - STAR: SQL Guided Pre-Training for Context-dependent Text-to-SQL Parsing [64.80483736666123]
We propose a novel pre-training framework STAR for context-dependent text-to- parsing.
In addition, we construct a large-scale context-dependent text-to-the-art conversation corpus to pre-train STAR.
Extensive experiments show that STAR achieves new state-of-the-art performance on two downstream benchmarks.
arXiv Detail & Related papers (2022-10-21T11:30:07Z) - A Survey on Text-to-SQL Parsing: Concepts, Methods, and Future
Directions [102.8606542189429]
The goal of text-to-corpora parsing is to convert a natural language (NL) question to its corresponding structured query language () based on the evidences provided by databases.
Deep neural networks have significantly advanced this task by neural generation models, which automatically learn a mapping function from an input NL question to an output query.
arXiv Detail & Related papers (2022-08-29T14:24:13Z) - Pay More Attention to History: A Context Modeling Strategy for
Conversational Text-to-SQL [8.038535788630542]
One of the most intractable problem of conversational text-to- domain is modeling the semantics of multi-turn queries.
This paper shows that explicit modeling the semantic changes by adding each turn and the summarization of the whole context can bring better performance.
arXiv Detail & Related papers (2021-12-16T09:41:04Z) - IGSQL: Database Schema Interaction Graph Based Neural Model for
Context-Dependent Text-to-SQL Generation [61.09660709356527]
We propose a database schema interaction graph encoder to utilize historicalal information of database schema items.
We evaluate our model on the benchmark SParC and Co datasets.
arXiv Detail & Related papers (2020-11-11T12:56:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.