Generating Synthetic Data for Task-Oriented Semantic Parsing with
Hierarchical Representations
- URL: http://arxiv.org/abs/2011.02050v1
- Date: Tue, 3 Nov 2020 22:55:40 GMT
- Title: Generating Synthetic Data for Task-Oriented Semantic Parsing with
Hierarchical Representations
- Authors: Ke Tran, Ming Tan
- Abstract summary: In this work, we explore the possibility of generating synthetic data for neural semantic parsing.
Specifically, we first extract masked templates from the existing labeled utterances, and then fine-tune BART to generate synthetic utterances conditioning.
We show the potential of our approach when evaluating on the Facebook TOP dataset for navigation domain.
- Score: 0.8203855808943658
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Modern conversational AI systems support natural language understanding for a
wide variety of capabilities. While a majority of these tasks can be
accomplished using a simple and flat representation of intents and slots, more
sophisticated capabilities require complex hierarchical representations
supported by semantic parsing. State-of-the-art semantic parsers are trained
using supervised learning with data labeled according to a hierarchical schema
which might be costly to obtain or not readily available for a new domain. In
this work, we explore the possibility of generating synthetic data for neural
semantic parsing using a pretrained denoising sequence-to-sequence model (i.e.,
BART). Specifically, we first extract masked templates from the existing
labeled utterances, and then fine-tune BART to generate synthetic utterances
conditioning on the extracted templates. Finally, we use an auxiliary parser
(AP) to filter the generated utterances. The AP guarantees the quality of the
generated data. We show the potential of our approach when evaluating on the
Facebook TOP dataset for navigation domain.
Related papers
- fPLSA: Learning Semantic Structures in Document Collections Using Foundation Models [19.099810900404357]
We introduce fPLSA, a foundation-model-based Probabilistic Latent Semantic Analysis (PLSA) method.
PLSA iteratively clusters and tags document segments based on document-level contexts.
Our experiments on story writing, math, and multi-step reasoning datasets demonstrate that fPLSA tags help reconstruct the original texts better than existing tagging methods.
arXiv Detail & Related papers (2024-10-07T20:25:52Z) - Synergizing Unsupervised and Supervised Learning: A Hybrid Approach for Accurate Natural Language Task Modeling [0.0]
This paper presents a novel hybrid approach that synergizes unsupervised and supervised learning to improve the accuracy of NLP task modeling.
Our methodology integrates an unsupervised module that learns representations from unlabeled corpora and a supervised module that leverages these representations to enhance task-specific models.
By synergizing techniques, our hybrid approach achieves SOTA results on benchmark datasets, paving the way for more data-efficient and robust NLP systems.
arXiv Detail & Related papers (2024-06-03T08:31:35Z) - Autoregressive Search Engines: Generating Substrings as Document
Identifiers [53.0729058170278]
Autoregressive language models are emerging as the de-facto standard for generating answers.
Previous work has explored ways to partition the search space into hierarchical structures.
In this work we propose an alternative that doesn't force any structure in the search space: using all ngrams in a passage as its possible identifiers.
arXiv Detail & Related papers (2022-04-22T10:45:01Z) - Hierarchical Neural Data Synthesis for Semantic Parsing [16.284764879030448]
We propose a purely neural approach of data augmentation for semantic parsing.
We achieve the state-of-the-art performance on the development set (77.2% accuracy) using our zero-shot augmentation.
arXiv Detail & Related papers (2021-12-04T01:33:08Z) - X2Parser: Cross-Lingual and Cross-Domain Framework for Task-Oriented
Compositional Semantic Parsing [51.81533991497547]
Task-oriented compositional semantic parsing (TCSP) handles complex nested user queries.
We present X2 compared a transferable Cross-lingual and Cross-domain for TCSP.
We propose to predict flattened intents and slots representations separately and cast both prediction tasks into sequence labeling problems.
arXiv Detail & Related papers (2021-06-07T16:40:05Z) - Low-Resource Task-Oriented Semantic Parsing via Intrinsic Modeling [65.51280121472146]
We exploit what we intrinsically know about ontology labels to build efficient semantic parsing models.
Our model is highly efficient using a low-resource benchmark derived from TOPv2.
arXiv Detail & Related papers (2021-04-15T04:01:02Z) - Learning to Synthesize Data for Semantic Parsing [57.190817162674875]
We propose a generative model which models the composition of programs and maps a program to an utterance.
Due to the simplicity of PCFG and pre-trained BART, our generative model can be efficiently learned from existing data at hand.
We evaluate our method in both in-domain and out-of-domain settings of text-to-Query parsing on the standard benchmarks of GeoQuery and Spider.
arXiv Detail & Related papers (2021-04-12T21:24:02Z) - Infusing Finetuning with Semantic Dependencies [62.37697048781823]
We show that, unlike syntax, semantics is not brought to the surface by today's pretrained models.
We then use convolutional graph encoders to explicitly incorporate semantic parses into task-specific finetuning.
arXiv Detail & Related papers (2020-12-10T01:27:24Z) - Unsupervised Paraphrasing with Pretrained Language Models [85.03373221588707]
We propose a training pipeline that enables pre-trained language models to generate high-quality paraphrases in an unsupervised setting.
Our recipe consists of task-adaptation, self-supervision, and a novel decoding algorithm named Dynamic Blocking.
We show with automatic and human evaluations that our approach achieves state-of-the-art performance on both the Quora Question Pair and the ParaNMT datasets.
arXiv Detail & Related papers (2020-10-24T11:55:28Z) - Assembling Semantically-Disentangled Representations for
Predictive-Generative Models via Adaptation from Synthetic Domain [32.42156485883356]
We show that semantically-aligned representations can be generated with the help of a physics based engine.
It is shown that the proposed (SYNTH-VAE-GAN) method can construct a conditional-generative model of human face attributes without relying on real data labels.
arXiv Detail & Related papers (2020-02-23T03:35:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.