Neural Bi-Lexicalized PCFG Induction
- URL: http://arxiv.org/abs/2105.15021v1
- Date: Mon, 31 May 2021 15:00:03 GMT
- Title: Neural Bi-Lexicalized PCFG Induction
- Authors: Songlin Yang, Yanpeng Zhao, Kewei Tu
- Abstract summary: We propose an approach to parameterize L-PCFGs without making implausible independence assumptions.
Our approach directly models bilexical dependencies and meanwhile reduces both learning and representation complexities of L-PCFGs.
- Score: 22.728124473130876
- License: http://creativecommons.org/publicdomain/zero/1.0/
- Abstract: Neural lexicalized PCFGs (L-PCFGs) have been shown effective in grammar
induction. However, to reduce computational complexity, they make a strong
independence assumption on the generation of the child word and thus bilexical
dependencies are ignored. In this paper, we propose an approach to parameterize
L-PCFGs without making implausible independence assumptions. Our approach
directly models bilexical dependencies and meanwhile reduces both learning and
representation complexities of L-PCFGs. Experimental results on the English WSJ
dataset confirm the effectiveness of our approach in improving both running
speed and unsupervised parsing performance.
Related papers
- Provenance: A Light-weight Fact-checker for Retrieval Augmented LLM Generation Output [49.893971654861424]
We present a light-weight approach for detecting nonfactual outputs from retrieval-augmented generation (RAG)
We compute a factuality score that can be thresholded to yield a binary decision.
Our experiments show high area under the ROC curve (AUC) across a wide range of relevant open source datasets.
arXiv Detail & Related papers (2024-11-01T20:44:59Z) - Investigating Critical Period Effects in Language Acquisition through Neural Language Models [70.6367059367609]
Second language (L2) acquisition becomes harder after early childhood.
ceasing exposure to a first language (L1) after this period (but not before) typically does not lead to substantial loss of L1 proficiency.
It is unknown whether these CP effects result from innately determined brain maturation or as a stabilization of neural connections naturally induced by experience.
arXiv Detail & Related papers (2024-07-27T19:17:10Z) - LINC: A Neurosymbolic Approach for Logical Reasoning by Combining
Language Models with First-Order Logic Provers [60.009969929857704]
Logical reasoning is an important task for artificial intelligence with potential impacts on science, mathematics, and society.
In this work, we reformulating such tasks as modular neurosymbolic programming, which we call LINC.
We observe significant performance gains on FOLIO and a balanced subset of ProofWriter for three different models in nearly all experimental conditions we evaluate.
arXiv Detail & Related papers (2023-10-23T17:58:40Z) - Simple Hardware-Efficient PCFGs with Independent Left and Right
Productions [77.12660133995362]
This work introduces emphSimplePCFG, a simple PCFG formalism with independent left and right productions.
As an unsupervised algorithm, our simple PCFG obtains an average F1 of 65.1 on the English PTB, and as a language model, it obtains a perplexity of 119.0, outperforming similarly-sized low-rank PCFGs.
arXiv Detail & Related papers (2023-10-23T14:48:51Z) - PCFGs Can Do Better: Inducing Probabilistic Context-Free Grammars with
Many Symbols [22.728124473130876]
We present a new parameterization form of PCFGs based on tensor decomposition.
We use neural parameterization for the new form to improve unsupervised parsing performance.
We evaluate our model across ten languages and empirically demonstrate the effectiveness of using more symbols.
arXiv Detail & Related papers (2021-04-28T12:25:27Z) - An Empirical Study of Compound PCFGs [35.64371385720051]
Compound context-free grammars (C-PCFGs) have recently established a new state of the art for unsupervised phrase-structure grammar induction.
We analyze C-PCFGs on English treebanks and conduct a multilingual evaluation of C-PCFGs.
The experimental results show that the best configurations of C-PCFGs, which are tuned on English, do not always generalize to morphology-rich languages.
arXiv Detail & Related papers (2021-03-03T10:24:26Z) - The Return of Lexical Dependencies: Neural Lexicalized PCFGs [103.41187595153652]
We present novel neural models of lexicalized PCFGs which allow us to overcome sparsity problems.
Experiments demonstrate that this unified framework results in stronger results on both representations than achieved when either formalism alone.
arXiv Detail & Related papers (2020-07-29T22:12:49Z) - Localized Debiased Machine Learning: Efficient Inference on Quantile
Treatment Effects and Beyond [69.83813153444115]
We consider an efficient estimating equation for the (local) quantile treatment effect ((L)QTE) in causal inference.
Debiased machine learning (DML) is a data-splitting approach to estimating high-dimensional nuisances.
We propose localized debiased machine learning (LDML), which avoids this burdensome step.
arXiv Detail & Related papers (2019-12-30T14:42:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.