Syntactic Substitutability as Unsupervised Dependency Syntax
- URL: http://arxiv.org/abs/2211.16031v3
- Date: Fri, 20 Oct 2023 18:10:57 GMT
- Title: Syntactic Substitutability as Unsupervised Dependency Syntax
- Authors: Jasper Jian and Siva Reddy
- Abstract summary: We model a more general property implicit in the definition of dependency relations, syntactic substitutability.
This property captures the fact that words at either end of a dependency can be substituted with words from the same category.
We show that increasing the number of substitutions used improves parsing accuracy on natural data.
- Score: 31.488677474152794
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Syntax is a latent hierarchical structure which underpins the robust and
compositional nature of human language. In this work, we explore the hypothesis
that syntactic dependencies can be represented in language model attention
distributions and propose a new method to induce these structures
theory-agnostically. Instead of modeling syntactic relations as defined by
annotation schemata, we model a more general property implicit in the
definition of dependency relations, syntactic substitutability. This property
captures the fact that words at either end of a dependency can be substituted
with words from the same category. Substitutions can be used to generate a set
of syntactically invariant sentences whose representations are then used for
parsing. We show that increasing the number of substitutions used improves
parsing accuracy on natural data. On long-distance subject-verb agreement
constructions, our method achieves 79.5% recall compared to 8.9% using a
previous method. Our method also provides improvements when transferred to a
different parsing setup, demonstrating that it generalizes.
Related papers
- Dynamic Syntax Mapping: A New Approach to Unsupervised Syntax Parsing [0.0]
This study investigates the premise that language models, specifically their attention distributions, can encapsulate syntactic dependencies.
We introduce Dynamic Syntax Mapping (DSM), an innovative approach for the induction of these structures.
Our findings reveal that the use of an increasing array of substitutions notably enhances parsing precision on natural language data.
arXiv Detail & Related papers (2023-12-18T10:34:29Z) - Conjunct Resolution in the Face of Verbal Omissions [51.220650412095665]
We propose a conjunct resolution task that operates directly on the text and makes use of a split-and-rephrase paradigm in order to recover the missing elements in the coordination structure.
We curate a large dataset, containing over 10K examples of naturally-occurring verbal omissions with crowd-sourced annotations.
We train various neural baselines for this task, and show that while our best method obtains decent performance, it leaves ample space for improvement.
arXiv Detail & Related papers (2023-05-26T08:44:02Z) - Structural Ambiguity and its Disambiguation in Language Model Based
Parsers: the Case of Dutch Clause Relativization [2.9950872478176627]
We study how the presence of a prior sentence can resolve relative clause ambiguities.
Results show that a neurosymbolic, based on proof nets, is more open to data bias correction than an approach based on universal dependencies.
arXiv Detail & Related papers (2023-05-24T09:04:18Z) - The distribution of syntactic dependency distances [0.7614628596146599]
We contribute to the characterization of the actual distribution of syntactic dependency distances.
We propose a new double-exponential model in which decay in probability is allowed to change after a break-point.
We find that a two-regime model is the most likely one in all 20 languages we considered.
arXiv Detail & Related papers (2022-11-26T17:31:25Z) - On The Ingredients of an Effective Zero-shot Semantic Parser [95.01623036661468]
We analyze zero-shot learning by paraphrasing training examples of canonical utterances and programs from a grammar.
We propose bridging these gaps using improved grammars, stronger paraphrasers, and efficient learning methods.
Our model achieves strong performance on two semantic parsing benchmarks (Scholar, Geo) with zero labeled data.
arXiv Detail & Related papers (2021-10-15T21:41:16Z) - Neural Abstructions: Abstractions that Support Construction for Grounded
Language Learning [69.1137074774244]
Leveraging language interactions effectively requires addressing limitations in the two most common approaches to language grounding.
We introduce the idea of neural abstructions: a set of constraints on the inference procedure of a label-conditioned generative model.
We show that with this method a user population is able to build a semantic modification for an open-ended house task in Minecraft.
arXiv Detail & Related papers (2021-07-20T07:01:15Z) - Syntactic Nuclei in Dependency Parsing -- A Multilingual Exploration [8.25332300240617]
We show how the concept of nucleus can be defined in the framework of Universal Dependencies.
Experiments on 12 languages show that nucleus composition gives small but significant improvements in parsing accuracy.
arXiv Detail & Related papers (2021-01-28T12:22:30Z) - Infusing Finetuning with Semantic Dependencies [62.37697048781823]
We show that, unlike syntax, semantics is not brought to the surface by today's pretrained models.
We then use convolutional graph encoders to explicitly incorporate semantic parses into task-specific finetuning.
arXiv Detail & Related papers (2020-12-10T01:27:24Z) - Unsupervised Distillation of Syntactic Information from Contextualized
Word Representations [62.230491683411536]
We tackle the task of unsupervised disentanglement between semantics and structure in neural language representations.
To this end, we automatically generate groups of sentences which are structurally similar but semantically different.
We demonstrate that our transformation clusters vectors in space by structural properties, rather than by lexical semantics.
arXiv Detail & Related papers (2020-10-11T15:13:18Z) - Transition-Based Dependency Parsing using Perceptron Learner [34.59241394911966]
We tackle transition-based dependency parsing using a Perceptron Learner.
Our proposed model, which adds more relevant features to the Perceptron Learner, outperforms a baseline arc-standard.
arXiv Detail & Related papers (2020-01-22T20:58:22Z) - Multiplex Word Embeddings for Selectional Preference Acquisition [70.33531759861111]
We propose a multiplex word embedding model, which can be easily extended according to various relations among words.
Our model can effectively distinguish words with respect to different relations without introducing unnecessary sparseness.
arXiv Detail & Related papers (2020-01-09T04:47:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.