mRAT-SQL+GAP:A Portuguese Text-to-SQL Transformer
- URL: http://arxiv.org/abs/2110.03546v1
- Date: Thu, 7 Oct 2021 15:08:24 GMT
- Title: mRAT-SQL+GAP:A Portuguese Text-to-SQL Transformer
- Authors: Marcelo Archanjo Jos\'e, Fabio Gagliardi Cozman
- Abstract summary: A large number of techniques are geared towards the English language.
In this work, we investigated translation tosql when input questions are given in a language different from English.
We changed the RAT-+GAP system by relying on a multilingual BART model.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: The translation of natural language questions to SQL queries has attracted
growing attention, in particular in connection with transformers and similar
language models. A large number of techniques are geared towards the English
language; in this work, we thus investigated translation to SQL when input
questions are given in the Portuguese language. To do so, we properly adapted
state-of-the-art tools and resources. We changed the RAT-SQL+GAP system by
relying on a multilingual BART model (we report tests with other language
models), and we produced a translated version of the Spider dataset. Our
experiments expose interesting phenomena that arise when non-English languages
are targeted; in particular, it is better to train with original and translated
training datasets together, even if a single target language is desired. This
multilingual BART model fine-tuned with a double-size training dataset (English
and Portuguese) achieved 83% of the baseline, making inferences for the
Portuguese test dataset. This investigation can help other researchers to
produce results in Machine Learning in a language different from English. Our
multilingual ready version of RAT-SQL+GAP and the data are available,
open-sourced as mRAT-SQL+GAP at: https://github.com/C4AI/gap-text2sql
Related papers
- Datasets for Multilingual Answer Sentence Selection [59.28492975191415]
We introduce new high-quality datasets for AS2 in five European languages (French, German, Italian, Portuguese, and Spanish)
Results indicate that our datasets are pivotal in producing robust and powerful multilingual AS2 models.
arXiv Detail & Related papers (2024-06-14T16:50:29Z) - StatBot.Swiss: Bilingual Open Data Exploration in Natural Language [5.149617340100317]
The potential for improvements brought by Large Language Models (LLMs) in Text-to-Swiss systems is mostly assessed on monolingual English datasets.
We release the StatBot, the first benchmark dataset for evaluating Text-to-Swiss systems based on real-world applications.
arXiv Detail & Related papers (2024-06-05T12:03:19Z) - Ar-Spider: Text-to-SQL in Arabic [11.463438573648297]
This paper introduces Ar-Spider 1, the first Arabic cross-language text-to-domain dataset.
Due to the unique nature of the language, two major challenges have been encountered, namely linguistic and structural challenges.
We propose the similarity relationship (CSR) approach, which results in a significant increase in the overall performance of about 1.52% for S2 and 1.06% for LGE and closes the gap between Arabic and English languages to 7.73%.
arXiv Detail & Related papers (2024-02-22T23:11:17Z) - Question Translation Training for Better Multilingual Reasoning [108.10066378240879]
Large language models show compelling performance on reasoning tasks but they tend to perform much worse in languages other than English.
A typical solution is to translate instruction data into all languages of interest, and then train on the resulting multilingual data, which is called translate-training.
In this paper we explore the benefits of question alignment, where we train the model to translate reasoning questions into English by finetuning on X-English parallel question data.
arXiv Detail & Related papers (2024-01-15T16:39:10Z) - Constructing Multilingual Code Search Dataset Using Neural Machine
Translation [48.32329232202801]
We create a multilingual code search dataset in four natural and four programming languages.
Our results show that the model pre-trained with all natural and programming language data has performed best in most cases.
arXiv Detail & Related papers (2023-06-27T16:42:36Z) - A Multilingual Translator to SQL with Database Schema Pruning to Improve
Self-Attention [0.0]
We present techniques that allow long text sequences to be handled by transformers with up to 512 input tokens.
In addition, we used a multilingual approach with the mT5-large model fine-tuned with a data-augmented Spider dataset in four languages simultaneously.
arXiv Detail & Related papers (2023-06-25T14:28:12Z) - MultiSpider: Towards Benchmarking Multilingual Text-to-SQL Semantic
Parsing [48.216386761482525]
We present MultiSpider, the largest multilingual text-to- schema- dataset which covers seven languages (English, German, French, Spanish, Japanese, Chinese, and Vietnamese)
Experimental results under three typical settings (zero-shot, monolingual and multilingual) reveal a 6.1% absolute drop in accuracy in non-English languages.
We also propose a simple framework augmentation framework SAVe (Augmentation-with-Verification) which boosts the overall performance by about 1.8% and closes the 29.5% performance gap across languages.
arXiv Detail & Related papers (2022-12-27T13:58:30Z) - XRICL: Cross-lingual Retrieval-Augmented In-Context Learning for
Cross-lingual Text-to-SQL Semantic Parsing [70.40401197026925]
In-context learning using large language models has recently shown surprising results for semantic parsing tasks.
This work introduces the XRICL framework, which learns to retrieve relevant English exemplars for a given query.
We also include global translation exemplars for a target language to facilitate the translation process for large language models.
arXiv Detail & Related papers (2022-10-25T01:33:49Z) - A Survey on Text-to-SQL Parsing: Concepts, Methods, and Future
Directions [102.8606542189429]
The goal of text-to-corpora parsing is to convert a natural language (NL) question to its corresponding structured query language () based on the evidences provided by databases.
Deep neural networks have significantly advanced this task by neural generation models, which automatically learn a mapping function from an input NL question to an output query.
arXiv Detail & Related papers (2022-08-29T14:24:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.