Learning grammar with a divide-and-concur neural network
- URL: http://arxiv.org/abs/2201.07341v1
- Date: Tue, 18 Jan 2022 22:42:43 GMT
- Title: Learning grammar with a divide-and-concur neural network
- Authors: Sean Deyo and Veit Elser
- Abstract summary: We implement a divide-and-concur iterative projection approach to context-free grammar inference.
Our method requires a relatively small number of discrete parameters, making the inferred grammar directly interpretable.
- Score: 4.111899441919164
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We implement a divide-and-concur iterative projection approach to
context-free grammar inference. Unlike most state-of-the-art models of natural
language processing, our method requires a relatively small number of discrete
parameters, making the inferred grammar directly interpretable -- one can read
off from a solution how to construct grammatically valid sentences. Another
advantage of our approach is the ability to infer meaningful grammatical rules
from just a few sentences, compared to the hundreds of gigabytes of training
data many other models employ. We demonstrate several ways of applying our
approach: classifying words and inferring a grammar from scratch, taking an
existing grammar and refining its categories and rules, and taking an existing
grammar and expanding its lexicon as it encounters new words in new data.
Related papers
- Leveraging Grammar Induction for Language Understanding and Generation [7.459693992079273]
We introduce an unsupervised grammar induction method for language understanding and generation.
We construct a grammar to induce constituency structures and dependency relations, which is simultaneously trained on downstream tasks.
We evaluate and apply our method to multiple machine translation tasks natural language understanding tasks.
arXiv Detail & Related papers (2024-10-07T09:57:59Z) - Incremental Context-free Grammar Inference in Black Box Settings [17.601446198181048]
Black-box context-free grammar inference is a significant challenge in many practical settings.
We propose a novel method that segments example strings into smaller units and incrementally infers the grammar.
Our approach, named Kedavra, has demonstrated superior grammar quality (enhanced precision and recall), faster runtime, and improved readability through empirical comparison.
arXiv Detail & Related papers (2024-08-29T17:00:38Z) - Detecting and explaining (in)equivalence of context-free grammars [0.6282171844772422]
We propose a scalable framework for deciding, proving, and explaining (in)equivalence of context-free grammars.
We present an implementation of the framework and evaluate it on large data sets collected within educational support systems.
arXiv Detail & Related papers (2024-07-25T17:36:18Z) - Language Models for Text Classification: Is In-Context Learning Enough? [54.869097980761595]
Recent foundational language models have shown state-of-the-art performance in many NLP tasks in zero- and few-shot settings.
An advantage of these models over more standard approaches is the ability to understand instructions written in natural language (prompts)
This makes them suitable for addressing text classification problems for domains with limited amounts of annotated instances.
arXiv Detail & Related papers (2024-03-26T12:47:39Z) - A logical word embedding for learning grammar [4.111899441919164]
We introduce the logical grammar emdebbing (LGE) to enable unsupervised inference of lexical categories and syntactic rules from a corpus of text.
LGE produces comprehensible output summarizing its inferences, has a completely transparent process for producing novel sentences, and can learn from as few as a hundred sentences.
arXiv Detail & Related papers (2023-04-28T01:53:54Z) - A Neural Model for Regular Grammar Induction [8.873449722727026]
We treat grammars as a model of computation and propose a novel neural approach to induction of regular grammars from positive and negative examples.
Our model is fully explainable, its intermediate results are directly interpretable as partial parses, and it can be used to learn arbitrary regular grammars when provided with sufficient data.
arXiv Detail & Related papers (2022-09-23T14:53:23Z) - Dependency Induction Through the Lens of Visual Perception [81.91502968815746]
We propose an unsupervised grammar induction model that leverages word concreteness and a structural vision-based to jointly learn constituency-structure and dependency-structure grammars.
Our experiments show that the proposed extension outperforms the current state-of-the-art visually grounded models in constituency parsing even with a smaller grammar size.
arXiv Detail & Related papers (2021-09-20T18:40:37Z) - Extracting Grammars from a Neural Network Parser for Anomaly Detection
in Unknown Formats [79.6676793507792]
Reinforcement learning has recently shown promise as a technique for training an artificial neural network to parse sentences in some unknown format.
This paper presents procedures for extracting production rules from the neural network, and for using these rules to determine whether a given sentence is nominal or anomalous.
arXiv Detail & Related papers (2021-07-30T23:10:24Z) - VLGrammar: Grounded Grammar Induction of Vision and Language [86.88273769411428]
We study grounded grammar induction of vision and language in a joint learning framework.
We present VLGrammar, a method that uses compound probabilistic context-free grammars (compound PCFGs) to induce the language grammar and the image grammar simultaneously.
arXiv Detail & Related papers (2021-03-24T04:05:08Z) - SLM: Learning a Discourse Language Representation with Sentence
Unshuffling [53.42814722621715]
We introduce Sentence-level Language Modeling, a new pre-training objective for learning a discourse language representation.
We show that this feature of our model improves the performance of the original BERT by large margins.
arXiv Detail & Related papers (2020-10-30T13:33:41Z) - Toward Better Storylines with Sentence-Level Language Models [54.91921545103256]
We propose a sentence-level language model which selects the next sentence in a story from a finite set of fluent alternatives.
We demonstrate the effectiveness of our approach with state-of-the-art accuracy on the unsupervised Story Cloze task.
arXiv Detail & Related papers (2020-05-11T16:54:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.