Conversational Text-to-SQL: An Odyssey into State-of-the-Art and
Challenges Ahead
- URL: http://arxiv.org/abs/2302.11054v1
- Date: Tue, 21 Feb 2023 23:15:33 GMT
- Title: Conversational Text-to-SQL: An Odyssey into State-of-the-Art and
Challenges Ahead
- Authors: Sree Hari Krishnan Parthasarathi, Lu Zeng, Dilek Hakkani-Tur
- Abstract summary: State-of-the-art (SOTA) systems use large, pre-trained and finetuned language models, such as the T5-family.
With multi-tasking (MT) over coherent tasks with discrete prompts during training, we improve over specialized text-to-three models.
We conduct studies to tease apart errors attributable to domain and compositional generalization.
- Score: 6.966624873109535
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Conversational, multi-turn, text-to-SQL (CoSQL) tasks map natural language
utterances in a dialogue to SQL queries. State-of-the-art (SOTA) systems use
large, pre-trained and finetuned language models, such as the T5-family, in
conjunction with constrained decoding. With multi-tasking (MT) over coherent
tasks with discrete prompts during training, we improve over specialized
text-to-SQL T5-family models. Based on Oracle analyses over n-best hypotheses,
we apply a query plan model and a schema linking algorithm as rerankers.
Combining MT and reranking, our results using T5-3B show absolute accuracy
improvements of 1.0% in exact match and 3.4% in execution match over a SOTA
baseline on CoSQL. While these gains consistently manifest at turn level,
context dependent turns are considerably harder. We conduct studies to tease
apart errors attributable to domain and compositional generalization, with the
latter remaining a challenge for multi-turn conversations, especially in
generating SQL with unseen parse trees.
Related papers
- SelECT-SQL: Self-correcting ensemble Chain-of-Thought for Text-to-SQL [3.422309388045878]
We introduce SelECT-, a novel in-context learning solution that uses an algorithmic combination of chain-of-thought, self-correction, and ensemble methods.
Specifically, when configured using GPT as the base LLM, SelECT-Turbo achieves 84.2% execution accuracy on the Spider leaderboard's development set.
arXiv Detail & Related papers (2024-09-16T05:40:18Z) - SQL-PaLM: Improved Large Language Model Adaptation for Text-to-SQL (extended) [53.95151604061761]
This paper introduces the framework for enhancing Text-to- filtering using large language models (LLMs)
With few-shot prompting, we explore the effectiveness of consistency decoding with execution-based error analyses.
With instruction fine-tuning, we delve deep in understanding the critical paradigms that influence the performance of tuned LLMs.
arXiv Detail & Related papers (2023-05-26T21:39:05Z) - UNITE: A Unified Benchmark for Text-to-SQL Evaluation [72.72040379293718]
We introduce a UNIfied benchmark for Text-to-domain systems.
It is composed of publicly available text-to-domain datasets and 29K databases.
Compared to the widely used Spider benchmark, we introduce a threefold increase in SQL patterns.
arXiv Detail & Related papers (2023-05-25T17:19:52Z) - Graphix-T5: Mixing Pre-Trained Transformers with Graph-Aware Layers for
Text-to-SQL Parsing [56.232873134174056]
One of the major challenges in text-to-text parsing is domain generalization, i.e., how to well generalize to unseen databases.
In this work, we explore ways to further augment the pre-trained text-to-text transformer model with specialized components for text-to-text parsing.
To this end, we propose a new architecture GRAPHIX-T5, augmented by some specially-designed graph-aware model with layers.
arXiv Detail & Related papers (2023-01-18T13:29:05Z) - MIGA: A Unified Multi-task Generation Framework for Conversational
Text-to-SQL [48.34333725045152]
Most state-of-the-art conversational text-to-generative methods are incompatible with pre-trained language models (PLMs), such as T5.
We present a two-stage unified MultI-task Generation frAmeme (MIGA) that leverages PLMs' ability to tackle conversational text-to-work.
arXiv Detail & Related papers (2022-12-19T07:14:32Z) - Towards Generalizable and Robust Text-to-SQL Parsing [77.18724939989647]
We propose a novel TKK framework consisting of Task decomposition, Knowledge acquisition, and Knowledge composition to learn text-to- parsing in stages.
We show that our framework is effective in all scenarios and state-of-the-art performance on the Spider, SParC, and Co. datasets.
arXiv Detail & Related papers (2022-10-23T09:21:27Z) - N-Best Hypotheses Reranking for Text-To-SQL Systems [6.966624873109535]
Text-to- task maps natural language utterances to structured queries.
State-of-the-art (SOTA) systems rely on finetuning large, pre-trained language models.
Findings show significant potential improvements with reranking.
arXiv Detail & Related papers (2022-10-19T15:35:06Z) - Weakly Supervised Text-to-SQL Parsing through Question Decomposition [53.22128541030441]
We take advantage of the recently proposed question meaning representation called QDMR.
Given questions, their QDMR structures (annotated by non-experts or automatically predicted) and the answers, we are able to automatically synthesizesql queries.
Our results show that the weakly supervised models perform competitively with those trained on NL- benchmark data.
arXiv Detail & Related papers (2021-12-12T20:02:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.