On The Ingredients of an Effective Zero-shot Semantic Parser
- URL: http://arxiv.org/abs/2110.08381v1
- Date: Fri, 15 Oct 2021 21:41:16 GMT
- Title: On The Ingredients of an Effective Zero-shot Semantic Parser
- Authors: Pengcheng Yin, John Wieting, Avirup Sil, Graham Neubig
- Abstract summary: We analyze zero-shot learning by paraphrasing training examples of canonical utterances and programs from a grammar.
We propose bridging these gaps using improved grammars, stronger paraphrasers, and efficient learning methods.
Our model achieves strong performance on two semantic parsing benchmarks (Scholar, Geo) with zero labeled data.
- Score: 95.01623036661468
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Semantic parsers map natural language utterances into meaning representations
(e.g., programs). Such models are typically bottlenecked by the paucity of
training data due to the required laborious annotation efforts. Recent studies
have performed zero-shot learning by synthesizing training examples of
canonical utterances and programs from a grammar, and further paraphrasing
these utterances to improve linguistic diversity. However, such synthetic
examples cannot fully capture patterns in real data. In this paper we analyze
zero-shot parsers through the lenses of the language and logical gaps (Herzig
and Berant, 2019), which quantify the discrepancy of language and programmatic
patterns between the canonical examples and real-world user-issued ones. We
propose bridging these gaps using improved grammars, stronger paraphrasers, and
efficient learning methods using canonical examples that most likely reflect
real user intents. Our model achieves strong performance on two semantic
parsing benchmarks (Scholar, Geo) with zero labeled data.
Related papers
- On Robustness of Prompt-based Semantic Parsing with Large Pre-trained
Language Model: An Empirical Study on Codex [48.588772371355816]
This paper presents the first empirical study on the adversarial robustness of a large prompt-based language model of code, codex.
Our results demonstrate that the state-of-the-art (SOTA) code-language models are vulnerable to carefully crafted adversarial examples.
arXiv Detail & Related papers (2023-01-30T13:21:00Z) - BenchCLAMP: A Benchmark for Evaluating Language Models on Syntactic and
Semantic Parsing [55.058258437125524]
We introduce BenchCLAMP, a Benchmark to evaluate Constrained LAnguage Model Parsing.
We benchmark eight language models, including two GPT-3 variants available only through an API.
Our experiments show that encoder-decoder pretrained language models can achieve similar performance or surpass state-of-the-art methods for syntactic and semantic parsing when the model output is constrained to be valid.
arXiv Detail & Related papers (2022-06-21T18:34:11Z) - Few-Shot Semantic Parsing with Language Models Trained On Code [52.23355024995237]
We find that Codex performs better at semantic parsing than equivalent GPT-3 models.
We find that unlike GPT-3, Codex performs similarly when targeting meaning representations directly, perhaps as meaning representations used in semantic parsing are structured similar to code.
arXiv Detail & Related papers (2021-12-16T08:34:06Z) - Constrained Language Models Yield Few-Shot Semantic Parsers [73.50960967598654]
We explore the use of large pretrained language models as few-shot semantics.
The goal in semantic parsing is to generate a structured meaning representation given a natural language input.
We use language models to paraphrase inputs into a controlled sublanguage resembling English that can be automatically mapped to a target meaning representation.
arXiv Detail & Related papers (2021-04-18T08:13:06Z) - Infusing Finetuning with Semantic Dependencies [62.37697048781823]
We show that, unlike syntax, semantics is not brought to the surface by today's pretrained models.
We then use convolutional graph encoders to explicitly incorporate semantic parses into task-specific finetuning.
arXiv Detail & Related papers (2020-12-10T01:27:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.