MIGA: A Unified Multi-task Generation Framework for Conversational
Text-to-SQL
- URL: http://arxiv.org/abs/2212.09278v1
- Date: Mon, 19 Dec 2022 07:14:32 GMT
- Title: MIGA: A Unified Multi-task Generation Framework for Conversational
Text-to-SQL
- Authors: Yingwen Fu, Wenjie Ou, Zhou Yu, and Yue Lin
- Abstract summary: Most state-of-the-art conversational text-to-generative methods are incompatible with pre-trained language models (PLMs), such as T5.
We present a two-stage unified MultI-task Generation frAmeme (MIGA) that leverages PLMs' ability to tackle conversational text-to-work.
- Score: 48.34333725045152
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Conversational text-to-SQL is designed to translate multi-turn natural
language questions into their corresponding SQL queries. Most state-of-the-art
conversational text- to-SQL methods are incompatible with generative
pre-trained language models (PLMs), such as T5. In this paper, we present a
two-stage unified MultI-task Generation frAmework (MIGA) that leverages PLMs'
ability to tackle conversational text-to-SQL. In the pre-training stage, MIGA
first decomposes the main task into several related sub-tasks and then unifies
them into the same sequence-to-sequence (Seq2Seq) paradigm with task-specific
natural language prompts to boost the main task from multi-task training. Later
in the fine-tuning stage, we propose four SQL perturbations to alleviate the
error propagation problem. MIGA tends to achieve state-of-the-art performance
on two benchmarks (SparC and CoSQL). We also provide extensive analyses and
discussions to shed light on some new perspectives for conversational
text-to-SQL.
Related papers
- QDA-SQL: Questions Enhanced Dialogue Augmentation for Multi-Turn Text-to-SQL [14.321009553155285]
Fine-tuned models often face challenges with multi-turn Text-to-answer tasks.
It is desired to enhance LLMs to handle multiple types of questions in multi-turn Text-to-answer tasks.
arXiv Detail & Related papers (2024-06-15T10:54:54Z) - Retrieval-augmented GPT-3.5-based Text-to-SQL Framework with
Sample-aware Prompting and Dynamic Revision Chain [21.593701177605652]
We propose a Text-to-aware prompting framework, involving a sample and a dynamic revision chain.
Our approach incorporates sample demonstrations and fine-grained information related to the given question.
To generate executable and accuratesqls without human intervention, we design a dynamic revision chain which iteratively adapts fine-grained feedback.
arXiv Detail & Related papers (2023-07-11T07:16:22Z) - SQL-PaLM: Improved Large Language Model Adaptation for Text-to-SQL (extended) [53.95151604061761]
This paper introduces the framework for enhancing Text-to- filtering using large language models (LLMs)
With few-shot prompting, we explore the effectiveness of consistency decoding with execution-based error analyses.
With instruction fine-tuning, we delve deep in understanding the critical paradigms that influence the performance of tuned LLMs.
arXiv Detail & Related papers (2023-05-26T21:39:05Z) - Divide and Prompt: Chain of Thought Prompting for Text-to-SQL [0.03807314298073299]
Chain-of-thought (CoT) prompting combined with large language models (LLMs) have achieved encouraging results on complex reasoning tasks.
We propose Divide-and-Prompt, which first divides the task into subtasks, and then approach each subtask through CoT.
arXiv Detail & Related papers (2023-04-23T06:52:35Z) - Conversational Text-to-SQL: An Odyssey into State-of-the-Art and
Challenges Ahead [6.966624873109535]
State-of-the-art (SOTA) systems use large, pre-trained and finetuned language models, such as the T5-family.
With multi-tasking (MT) over coherent tasks with discrete prompts during training, we improve over specialized text-to-three models.
We conduct studies to tease apart errors attributable to domain and compositional generalization.
arXiv Detail & Related papers (2023-02-21T23:15:33Z) - Towards Generalizable and Robust Text-to-SQL Parsing [77.18724939989647]
We propose a novel TKK framework consisting of Task decomposition, Knowledge acquisition, and Knowledge composition to learn text-to- parsing in stages.
We show that our framework is effective in all scenarios and state-of-the-art performance on the Spider, SParC, and Co. datasets.
arXiv Detail & Related papers (2022-10-23T09:21:27Z) - A Survey on Text-to-SQL Parsing: Concepts, Methods, and Future
Directions [102.8606542189429]
The goal of text-to-corpora parsing is to convert a natural language (NL) question to its corresponding structured query language () based on the evidences provided by databases.
Deep neural networks have significantly advanced this task by neural generation models, which automatically learn a mapping function from an input NL question to an output query.
arXiv Detail & Related papers (2022-08-29T14:24:13Z) - Bridging Cross-Lingual Gaps During Leveraging the Multilingual
Sequence-to-Sequence Pretraining for Text Generation [80.16548523140025]
We extend the vanilla pretrain-finetune pipeline with extra code-switching restore task to bridge the gap between the pretrain and finetune stages.
Our approach could narrow the cross-lingual sentence representation distance and improve low-frequency word translation with trivial computational cost.
arXiv Detail & Related papers (2022-04-16T16:08:38Z) - Pay More Attention to History: A Context Modeling Strategy for
Conversational Text-to-SQL [8.038535788630542]
One of the most intractable problem of conversational text-to- domain is modeling the semantics of multi-turn queries.
This paper shows that explicit modeling the semantic changes by adding each turn and the summarization of the whole context can bring better performance.
arXiv Detail & Related papers (2021-12-16T09:41:04Z) - Weakly Supervised Text-to-SQL Parsing through Question Decomposition [53.22128541030441]
We take advantage of the recently proposed question meaning representation called QDMR.
Given questions, their QDMR structures (annotated by non-experts or automatically predicted) and the answers, we are able to automatically synthesizesql queries.
Our results show that the weakly supervised models perform competitively with those trained on NL- benchmark data.
arXiv Detail & Related papers (2021-12-12T20:02:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.