The Return of Lexical Dependencies: Neural Lexicalized PCFGs
- URL: http://arxiv.org/abs/2007.15135v1
- Date: Wed, 29 Jul 2020 22:12:49 GMT
- Title: The Return of Lexical Dependencies: Neural Lexicalized PCFGs
- Authors: Hao Zhu, Yonatan Bisk, Graham Neubig
- Abstract summary: We present novel neural models of lexicalized PCFGs which allow us to overcome sparsity problems.
Experiments demonstrate that this unified framework results in stronger results on both representations than achieved when either formalism alone.
- Score: 103.41187595153652
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: In this paper we demonstrate that $\textit{context free grammar (CFG) based
methods for grammar induction benefit from modeling lexical dependencies}$.
This contrasts to the most popular current methods for grammar induction, which
focus on discovering $\textit{either}$ constituents $\textit{or}$ dependencies.
Previous approaches to marry these two disparate syntactic formalisms (e.g.
lexicalized PCFGs) have been plagued by sparsity, making them unsuitable for
unsupervised grammar induction. However, in this work, we present novel neural
models of lexicalized PCFGs which allow us to overcome sparsity problems and
effectively induce both constituents and dependencies within a single model.
Experiments demonstrate that this unified framework results in stronger results
on both representations than achieved when modeling either formalism alone.
Code is available at https://github.com/neulab/neural-lpcfg.
Related papers
- LINC: A Neurosymbolic Approach for Logical Reasoning by Combining
Language Models with First-Order Logic Provers [60.009969929857704]
Logical reasoning is an important task for artificial intelligence with potential impacts on science, mathematics, and society.
In this work, we reformulating such tasks as modular neurosymbolic programming, which we call LINC.
We observe significant performance gains on FOLIO and a balanced subset of ProofWriter for three different models in nearly all experimental conditions we evaluate.
arXiv Detail & Related papers (2023-10-23T17:58:40Z) - Transparency at the Source: Evaluating and Interpreting Language Models
With Access to the True Distribution [4.01799362940916]
We present a setup for training, evaluating and interpreting neural language models, that uses artificial, language-like data.
The data is generated using a massive probabilistic grammar, that is itself derived from a large natural language corpus.
With access to the underlying true source, our results show striking differences and outcomes in learning dynamics between different classes of words.
arXiv Detail & Related papers (2023-10-23T12:03:01Z) - Dependency Induction Through the Lens of Visual Perception [81.91502968815746]
We propose an unsupervised grammar induction model that leverages word concreteness and a structural vision-based to jointly learn constituency-structure and dependency-structure grammars.
Our experiments show that the proposed extension outperforms the current state-of-the-art visually grounded models in constituency parsing even with a smaller grammar size.
arXiv Detail & Related papers (2021-09-20T18:40:37Z) - Extracting Grammars from a Neural Network Parser for Anomaly Detection
in Unknown Formats [79.6676793507792]
Reinforcement learning has recently shown promise as a technique for training an artificial neural network to parse sentences in some unknown format.
This paper presents procedures for extracting production rules from the neural network, and for using these rules to determine whether a given sentence is nominal or anomalous.
arXiv Detail & Related papers (2021-07-30T23:10:24Z) - The Limitations of Limited Context for Constituency Parsing [27.271792317099045]
Parsing-Reading-Predict architecture of (Shen et al., 2018a) was first to perform unsupervised syntactic parsing.
What kind of syntactic structure can current neural approaches to syntax represent?
We ground this question in the sandbox of probabilistic-free-grammars (PCFGs)
We identify a key aspect of the representational power of these approaches: the amount and directionality of context that the predictor has access to.
arXiv Detail & Related papers (2021-06-03T03:58:35Z) - Neural Bi-Lexicalized PCFG Induction [22.728124473130876]
We propose an approach to parameterize L-PCFGs without making implausible independence assumptions.
Our approach directly models bilexical dependencies and meanwhile reduces both learning and representation complexities of L-PCFGs.
arXiv Detail & Related papers (2021-05-31T15:00:03Z) - Rule Augmented Unsupervised Constituency Parsing [11.775897250472116]
We propose an approach that utilizes very generic linguistic knowledge of the language present in the form of syntactic rules.
We achieve new state-of-the-art results on two benchmarks datasets, MNLI and WSJ.
arXiv Detail & Related papers (2021-05-21T08:06:11Z) - Lexically-constrained Text Generation through Commonsense Knowledge
Extraction and Injection [62.071938098215085]
We focus on the Commongen benchmark, wherein the aim is to generate a plausible sentence for a given set of input concepts.
We propose strategies for enhancing the semantic correctness of the generated text.
arXiv Detail & Related papers (2020-12-19T23:23:40Z) - Second-Order Unsupervised Neural Dependency Parsing [52.331561380948564]
Most unsupervised dependencys are based on first-order probabilistic generative models that only consider local parent-child information.
Inspired by second-order supervised dependency parsing, we proposed a second-order extension of unsupervised neural dependency models that incorporate grandparent-child or sibling information.
Our joint model achieves a 10% improvement over the previous state-of-the-art on the full WSJ test set.
arXiv Detail & Related papers (2020-10-28T03:01:33Z) - Traduction des Grammaires Cat\'egorielles de Lambek dans les Grammaires
Cat\'egorielles Abstraites [0.0]
This internship report is to demonstrate that every Lambek Grammar can be, not entirely but efficiently, expressed in Abstract Categorial Grammars (ACG)
The main idea is to transform the type rewriting system of LGs into that of Context-Free Grammars (CFG) by erasing introduction and elimination rules and generating enough axioms so that the cut rule suffices.
Although the underlying algorithm was not fully implemented, this proof provides another argument in favour of the relevance of ACGs in Natural Language Processing.
arXiv Detail & Related papers (2020-01-23T18:23:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.