Relation Aware Semi-autoregressive Semantic Parsing for NL2SQL
- URL: http://arxiv.org/abs/2108.00804v1
- Date: Mon, 2 Aug 2021 12:21:08 GMT
- Title: Relation Aware Semi-autoregressive Semantic Parsing for NL2SQL
- Authors: Junyang Huang, Yongbo Wang, Yongliang Wang, Yang Dong and Yanghua Xiao
- Abstract summary: We present a Relation aware Semi-autogressive Semantic Parsing (MODN) framework, which is more adaptable for NL2 backbone.
From empirical results and case study, our model shows its effectiveness in learning better word representation in NL2.
- Score: 17.605904256822786
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Natural language to SQL (NL2SQL) aims to parse a natural language with a
given database into a SQL query, which widely appears in practical Internet
applications. Jointly encode database schema and question utterance is a
difficult but important task in NL2SQL. One solution is to treat the input as a
heterogeneous graph. However, it failed to learn good word representation in
question utterance. Learning better word representation is important for
constructing a well-designed NL2SQL system. To solve the challenging task, we
present a Relation aware Semi-autogressive Semantic Parsing (\MODN) ~framework,
which is more adaptable for NL2SQL. It first learns relation embedding over the
schema entities and question words with predefined schema relations with
ELECTRA and relation aware transformer layer as backbone. Then we decode the
query SQL with a semi-autoregressive parser and predefined SQL syntax. From
empirical results and case study, our model shows its effectiveness in learning
better word representation in NL2SQL.
Related papers
- UNITE: A Unified Benchmark for Text-to-SQL Evaluation [72.72040379293718]
We introduce a UNIfied benchmark for Text-to-domain systems.
It is composed of publicly available text-to-domain datasets and 29K databases.
Compared to the widely used Spider benchmark, we introduce a threefold increase in SQL patterns.
arXiv Detail & Related papers (2023-05-25T17:19:52Z) - Prompting GPT-3.5 for Text-to-SQL with De-semanticization and Skeleton
Retrieval [17.747079214502673]
Text-to- is a task that converts a natural language question into a structured query language () to retrieve information from a database.
In this paper, we propose an LLM-based framework for Text-to- which retrieves helpful demonstration examples to prompt LLMs.
We design a de-semanticization mechanism that extracts question skeletons, allowing us to retrieve similar examples based on their structural similarity.
arXiv Detail & Related papers (2023-04-26T06:02:01Z) - XRICL: Cross-lingual Retrieval-Augmented In-Context Learning for
Cross-lingual Text-to-SQL Semantic Parsing [70.40401197026925]
In-context learning using large language models has recently shown surprising results for semantic parsing tasks.
This work introduces the XRICL framework, which learns to retrieve relevant English exemplars for a given query.
We also include global translation exemplars for a target language to facilitate the translation process for large language models.
arXiv Detail & Related papers (2022-10-25T01:33:49Z) - Improving Text-to-SQL Semantic Parsing with Fine-grained Query
Understanding [84.04706075621013]
We present a general-purpose, modular neural semantic parsing framework based on token-level fine-grained query understanding.
Our framework consists of three modules: named entity recognizer (NER), neural entity linker (NEL) and neural entity linker (NSP)
arXiv Detail & Related papers (2022-09-28T21:00:30Z) - A Survey on Text-to-SQL Parsing: Concepts, Methods, and Future
Directions [102.8606542189429]
The goal of text-to-corpora parsing is to convert a natural language (NL) question to its corresponding structured query language () based on the evidences provided by databases.
Deep neural networks have significantly advanced this task by neural generation models, which automatically learn a mapping function from an input NL question to an output query.
arXiv Detail & Related papers (2022-08-29T14:24:13Z) - HIE-SQL: History Information Enhanced Network for Context-Dependent
Text-to-SQL Semantic Parsing [1.343950231082215]
We propose a History Information Enhanced text-to-the-art model (HIE-) to exploit context-dependence information from both history utterances and the last predictedsql query.
We show our methods improve the performance of HIE- by a significant margin, which achieves new state-of-the-art results on the two context-dependent text-to-the-art benchmarks.
arXiv Detail & Related papers (2022-03-14T11:58:37Z) - S$^2$SQL: Injecting Syntax to Question-Schema Interaction Graph Encoder
for Text-to-SQL Parsers [66.78665327694625]
We propose S$2$, injecting Syntax to question- encoder graph for Text-to- relational parsing.
We also employ the decoupling constraint to induce diverse edge embedding, which further improves the network's performance.
Experiments on the Spider and robustness setting Spider-Syn demonstrate that the proposed approach outperforms all existing methods when pre-training models are used.
arXiv Detail & Related papers (2022-03-14T09:49:15Z) - Weakly Supervised Text-to-SQL Parsing through Question Decomposition [53.22128541030441]
We take advantage of the recently proposed question meaning representation called QDMR.
Given questions, their QDMR structures (annotated by non-experts or automatically predicted) and the answers, we are able to automatically synthesizesql queries.
Our results show that the weakly supervised models perform competitively with those trained on NL- benchmark data.
arXiv Detail & Related papers (2021-12-12T20:02:42Z) - ShadowGNN: Graph Projection Neural Network for Text-to-SQL Parser [36.12921337235763]
We propose a new architecture, ShadowGNN, which processes schemas at abstract and semantic levels.
On the challenging Text-to-Spider benchmark, empirical results show that ShadowGNN outperforms state-of-the-art models.
arXiv Detail & Related papers (2021-04-10T05:48:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.