Rule Augmented Unsupervised Constituency Parsing
- URL: http://arxiv.org/abs/2105.10193v1
- Date: Fri, 21 May 2021 08:06:11 GMT
- Title: Rule Augmented Unsupervised Constituency Parsing
- Authors: Atul Sahay, Anshul Nasery, Ayush Maheshwari, Ganesh Ramakrishnan and
Rishabh Iyer
- Abstract summary: We propose an approach that utilizes very generic linguistic knowledge of the language present in the form of syntactic rules.
We achieve new state-of-the-art results on two benchmarks datasets, MNLI and WSJ.
- Score: 11.775897250472116
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recently, unsupervised parsing of syntactic trees has gained considerable
attention. A prototypical approach to such unsupervised parsing employs
reinforcement learning and auto-encoders. However, no mechanism ensures that
the learnt model leverages the well-understood language grammar. We propose an
approach that utilizes very generic linguistic knowledge of the language
present in the form of syntactic rules, thus inducing better syntactic
structures. We introduce a novel formulation that takes advantage of the
syntactic grammar rules and is independent of the base system. We achieve new
state-of-the-art results on two benchmarks datasets, MNLI and WSJ. The source
code of the paper is available at https://github.com/anshuln/Diora_with_rules.
Related papers
- On Eliciting Syntax from Language Models via Hashing [19.872554909401316]
Unsupervised parsing aims to infer syntactic structure from raw text.
In this paper, we explore the possibility of leveraging this capability to deduce parsing trees from raw text.
We show that our method is effective and efficient enough to acquire high-quality parsing trees from pre-trained language models at a low cost.
arXiv Detail & Related papers (2024-10-05T08:06:19Z) - On Unsupervised Training of Link Grammar Based Language Models [0.0]
We introduce the ter-mination tags formalism required to build a language model based on a link grammar formalism.
Second, we pro-pose a statistical link grammar formalism, allowing for statistical language generation.
Third, based on the above formalism, we show that the classical dissertation of Yuret [25] on discovery of linguistic relations using lexical at-traction ignores contextual properties of the language.
arXiv Detail & Related papers (2022-08-27T14:07:24Z) - Incorporating Constituent Syntax for Coreference Resolution [50.71868417008133]
We propose a graph-based method to incorporate constituent syntactic structures.
We also explore to utilise higher-order neighbourhood information to encode rich structures in constituent trees.
Experiments on the English and Chinese portions of OntoNotes 5.0 benchmark show that our proposed model either beats a strong baseline or achieves new state-of-the-art performance.
arXiv Detail & Related papers (2022-02-22T07:40:42Z) - On The Ingredients of an Effective Zero-shot Semantic Parser [95.01623036661468]
We analyze zero-shot learning by paraphrasing training examples of canonical utterances and programs from a grammar.
We propose bridging these gaps using improved grammars, stronger paraphrasers, and efficient learning methods.
Our model achieves strong performance on two semantic parsing benchmarks (Scholar, Geo) with zero labeled data.
arXiv Detail & Related papers (2021-10-15T21:41:16Z) - Dependency Induction Through the Lens of Visual Perception [81.91502968815746]
We propose an unsupervised grammar induction model that leverages word concreteness and a structural vision-based to jointly learn constituency-structure and dependency-structure grammars.
Our experiments show that the proposed extension outperforms the current state-of-the-art visually grounded models in constituency parsing even with a smaller grammar size.
arXiv Detail & Related papers (2021-09-20T18:40:37Z) - Infusing Finetuning with Semantic Dependencies [62.37697048781823]
We show that, unlike syntax, semantics is not brought to the surface by today's pretrained models.
We then use convolutional graph encoders to explicitly incorporate semantic parses into task-specific finetuning.
arXiv Detail & Related papers (2020-12-10T01:27:24Z) - Automatic Extraction of Rules Governing Morphological Agreement [103.78033184221373]
We develop an automated framework for extracting a first-pass grammatical specification from raw text.
We focus on extracting rules describing agreement, a morphosyntactic phenomenon at the core of the grammars of many of the world's languages.
We apply our framework to all languages included in the Universal Dependencies project, with promising results.
arXiv Detail & Related papers (2020-10-02T18:31:45Z) - The Return of Lexical Dependencies: Neural Lexicalized PCFGs [103.41187595153652]
We present novel neural models of lexicalized PCFGs which allow us to overcome sparsity problems.
Experiments demonstrate that this unified framework results in stronger results on both representations than achieved when either formalism alone.
arXiv Detail & Related papers (2020-07-29T22:12:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.