Pay More Attention to History: A Context Modeling Strategy for
Conversational Text-to-SQL
- URL: http://arxiv.org/abs/2112.08735v1
- Date: Thu, 16 Dec 2021 09:41:04 GMT
- Title: Pay More Attention to History: A Context Modeling Strategy for
Conversational Text-to-SQL
- Authors: Yuntao Li, Hanchu Zhang, Yutian Li, Sirui Wang, Wei Wu, Yan Zhang
- Abstract summary: One of the most intractable problem of conversational text-to- domain is modeling the semantics of multi-turn queries.
This paper shows that explicit modeling the semantic changes by adding each turn and the summarization of the whole context can bring better performance.
- Score: 8.038535788630542
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Conversational text-to-SQL aims at converting multi-turn natural language
queries into their corresponding SQL representations. One of the most
intractable problem of conversational text-to-SQL is modeling the semantics of
multi-turn queries and gathering proper information required for the current
query. This paper shows that explicit modeling the semantic changes by adding
each turn and the summarization of the whole context can bring better
performance on converting conversational queries into SQLs. In particular, we
propose two conversational modeling tasks in both turn grain and conversation
grain. These two tasks simply work as auxiliary training tasks to help with
multi-turn conversational semantic parsing. We conducted empirical studies and
achieve new state-of-the-art results on large-scale open-domain conversational
text-to-SQL dataset. The results demonstrate that the proposed mechanism
significantly improves the performance of multi-turn semantic parsing.
Related papers
- SQL-PaLM: Improved Large Language Model Adaptation for Text-to-SQL (extended) [53.95151604061761]
This paper introduces the framework for enhancing Text-to- filtering using large language models (LLMs)
With few-shot prompting, we explore the effectiveness of consistency decoding with execution-based error analyses.
With instruction fine-tuning, we delve deep in understanding the critical paradigms that influence the performance of tuned LLMs.
arXiv Detail & Related papers (2023-05-26T21:39:05Z) - MIGA: A Unified Multi-task Generation Framework for Conversational
Text-to-SQL [48.34333725045152]
Most state-of-the-art conversational text-to-generative methods are incompatible with pre-trained language models (PLMs), such as T5.
We present a two-stage unified MultI-task Generation frAmeme (MIGA) that leverages PLMs' ability to tackle conversational text-to-work.
arXiv Detail & Related papers (2022-12-19T07:14:32Z) - Augmenting Multi-Turn Text-to-SQL Datasets with Self-Play [46.07002748587857]
We explore augmenting the training datasets using self-play, which leverages contextual information to synthesize new interactions.
We find that self-play improves the accuracy of a strong baseline on SParC and Co, two widely used text-to-domain datasets.
arXiv Detail & Related papers (2022-10-21T16:40:07Z) - STAR: SQL Guided Pre-Training for Context-dependent Text-to-SQL Parsing [64.80483736666123]
We propose a novel pre-training framework STAR for context-dependent text-to- parsing.
In addition, we construct a large-scale context-dependent text-to-the-art conversation corpus to pre-train STAR.
Extensive experiments show that STAR achieves new state-of-the-art performance on two downstream benchmarks.
arXiv Detail & Related papers (2022-10-21T11:30:07Z) - A Survey on Text-to-SQL Parsing: Concepts, Methods, and Future
Directions [102.8606542189429]
The goal of text-to-corpora parsing is to convert a natural language (NL) question to its corresponding structured query language () based on the evidences provided by databases.
Deep neural networks have significantly advanced this task by neural generation models, which automatically learn a mapping function from an input NL question to an output query.
arXiv Detail & Related papers (2022-08-29T14:24:13Z) - S$^2$SQL: Injecting Syntax to Question-Schema Interaction Graph Encoder
for Text-to-SQL Parsers [66.78665327694625]
We propose S$2$, injecting Syntax to question- encoder graph for Text-to- relational parsing.
We also employ the decoupling constraint to induce diverse edge embedding, which further improves the network's performance.
Experiments on the Spider and robustness setting Spider-Syn demonstrate that the proposed approach outperforms all existing methods when pre-training models are used.
arXiv Detail & Related papers (2022-03-14T09:49:15Z) - Decoupled Dialogue Modeling and Semantic Parsing for Multi-Turn
Text-to-SQL [20.92732277474218]
We propose a novel decoupled multi-turn Text-to-end framework, where an utterance rewrite model first explicitly solves completion of dialogue context.
A dual learning approach is also proposed for the utterance rewrite model to address the data sparsity problem.
With just a few rewrite cases, the decoupled method outperforms the released state-of-the-art end-to-end models on both SParC and Co datasets.
arXiv Detail & Related papers (2021-06-04T06:31:39Z) - Tracking Interaction States for Multi-Turn Text-to-SQL Semantic Parsing [44.0348697408427]
The task of multi-turn text-to- semantic parsing aims to translate natural language utterances in an interaction intosql queries.
A graph relational network and a non-linear layer are designed to update the representations of these two states respectively.
Experimental results on the challenging Co dataset demonstrate the effectiveness of our proposed method.
arXiv Detail & Related papers (2020-12-09T11:59:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.