RL-GRIT: Reinforcement Learning for Grammar Inference
- URL: http://arxiv.org/abs/2105.13114v1
- Date: Mon, 17 May 2021 23:48:39 GMT
- Title: RL-GRIT: Reinforcement Learning for Grammar Inference
- Authors: Walt Woods
- Abstract summary: We propose a novel set of mechanisms for grammar inference, RL-GRIT, and apply them to understanding de facto data formats.
Within this work, we lay out the many algorithmic changes required to adapt RL from its traditional, sequential-time environment to the highly interdependent environment of parsing.
- Score: 2.741266294612776
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: When working to understand usage of a data format, examples of the data
format are often more representative than the format's specification. For
example, two different applications might use very different JSON
representations, or two PDF-writing applications might make use of very
different areas of the PDF specification to realize the same rendered content.
The complexity arising from these distinct origins can lead to large,
difficult-to-understand attack surfaces, presenting a security concern when
considering both exfiltration and data schizophrenia. Grammar inference can aid
in describing the practical language generator behind examples of a data
format. However, most grammar inference research focuses on natural language,
not data formats, and fails to support crucial features such as type recursion.
We propose a novel set of mechanisms for grammar inference, RL-GRIT, and apply
them to understanding de facto data formats. After reviewing existing grammar
inference solutions, it was determined that a new, more flexible scaffold could
be found in Reinforcement Learning (RL). Within this work, we lay out the many
algorithmic changes required to adapt RL from its traditional, sequential-time
environment to the highly interdependent environment of parsing. The result is
an algorithm which can demonstrably learn recursive control structures in
simple data formats, and can extract meaningful structure from fragments of the
PDF format. Whereas prior work in grammar inference focused on either regular
languages or constituency parsing, we show that RL can be used to surpass the
expressiveness of both classes, and offers a clear path to learning
context-sensitive languages. The proposed algorithm can serve as a building
block for understanding the ecosystems of de facto data formats.
Related papers
- Language Models for Text Classification: Is In-Context Learning Enough? [54.869097980761595]
Recent foundational language models have shown state-of-the-art performance in many NLP tasks in zero- and few-shot settings.
An advantage of these models over more standard approaches is the ability to understand instructions written in natural language (prompts)
This makes them suitable for addressing text classification problems for domains with limited amounts of annotated instances.
arXiv Detail & Related papers (2024-03-26T12:47:39Z) - Soft Language Clustering for Multilingual Model Pre-training [57.18058739931463]
We propose XLM-P, which contextually retrieves prompts as flexible guidance for encoding instances conditionally.
Our XLM-P enables (1) lightweight modeling of language-invariant and language-specific knowledge across languages, and (2) easy integration with other multilingual pre-training methods.
arXiv Detail & Related papers (2023-06-13T08:08:08Z) - Physics of Language Models: Part 1, Learning Hierarchical Language Structures [51.68385617116854]
Transformer-based language models are effective but complex, and understanding their inner workings is a significant challenge.
We introduce a family of synthetic CFGs that produce hierarchical rules, capable of generating lengthy sentences.
We demonstrate that generative models like GPT can accurately learn this CFG language and generate sentences based on it.
arXiv Detail & Related papers (2023-05-23T04:28:16Z) - Unified Text Structuralization with Instruction-tuned Language Models [28.869098023025753]
We propose a simple and efficient approach to instruct large language model (LLM) to extract a variety of structures from texts.
Experiments show that this approach can enable language models to perform comparable with other state-of-the-art methods on datasets of a variety of languages and knowledge.
arXiv Detail & Related papers (2023-03-27T07:39:05Z) - MURMUR: Modular Multi-Step Reasoning for Semi-Structured Data-to-Text
Generation [102.20036684996248]
We propose MURMUR, a neuro-symbolic modular approach to text generation from semi-structured data with multi-step reasoning.
We conduct experiments on two data-to-text generation tasks like WebNLG and LogicNLG.
arXiv Detail & Related papers (2022-12-16T17:36:23Z) - UnifieR: A Unified Retriever for Large-Scale Retrieval [84.61239936314597]
Large-scale retrieval is to recall relevant documents from a huge collection given a query.
Recent retrieval methods based on pre-trained language models (PLM) can be coarsely categorized into either dense-vector or lexicon-based paradigms.
We propose a new learning framework, UnifieR which unifies dense-vector and lexicon-based retrieval in one model with a dual-representing capability.
arXiv Detail & Related papers (2022-05-23T11:01:59Z) - On The Ingredients of an Effective Zero-shot Semantic Parser [95.01623036661468]
We analyze zero-shot learning by paraphrasing training examples of canonical utterances and programs from a grammar.
We propose bridging these gaps using improved grammars, stronger paraphrasers, and efficient learning methods.
Our model achieves strong performance on two semantic parsing benchmarks (Scholar, Geo) with zero labeled data.
arXiv Detail & Related papers (2021-10-15T21:41:16Z) - Cross-Lingual Adaptation for Type Inference [29.234418962960905]
We propose a cross-lingual adaptation framework, PLATO, to transfer a deep learning-based type inference procedure across weakly typed languages.
By leveraging data from strongly typed languages, PLATO improves the perplexity of the backbone cross-programming-language model.
arXiv Detail & Related papers (2021-07-01T00:20:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.