SPAWNing Structural Priming Predictions from a Cognitively Motivated
Parser
- URL: http://arxiv.org/abs/2403.07202v1
- Date: Mon, 11 Mar 2024 22:58:58 GMT
- Title: SPAWNing Structural Priming Predictions from a Cognitively Motivated
Parser
- Authors: Grusha Prasad and Tal Linzen
- Abstract summary: We propose a framework for using empirical priming patterns to build a theory characterizing the structural representations humans construct when processing sentences.
We use SPAWN to generate priming predictions from two theoretical accounts which make different assumptions about the structure of relative clauses.
We find that the predictions from only one of these theories align with empirical priming patterns, thus highlighting which assumptions about relative clause better capture human sentence representations.
- Score: 24.613169009428827
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Structural priming is a widely used psycholinguistic paradigm to study human
sentence representations. In this work we propose a framework for using
empirical priming patterns to build a theory characterizing the structural
representations humans construct when processing sentences. This framework uses
a new cognitively motivated parser, SPAWN, to generate quantitative priming
predictions from theoretical syntax and evaluate these predictions with
empirical human behavior. As a case study, we apply this framework to study
reduced relative clause representations in English. We use SPAWN to generate
priming predictions from two theoretical accounts which make different
assumptions about the structure of relative clauses. We find that the
predictions from only one of these theories (Participial-Phase) align with
empirical priming patterns, thus highlighting which assumptions about relative
clause better capture human sentence representations.
Related papers
- The Foundations of Tokenization: Statistical and Computational Concerns [51.370165245628975]
Tokenization is a critical step in the NLP pipeline.
Despite its recognized importance as a standard representation method in NLP, the theoretical underpinnings of tokenization are not yet fully understood.
The present paper contributes to addressing this theoretical gap by proposing a unified formal framework for representing and analyzing tokenizer models.
arXiv Detail & Related papers (2024-07-16T11:12:28Z) - eRST: A Signaled Graph Theory of Discourse Relations and Organization [14.074017875514787]
We present a new theoretical framework for computational discourse analysis, based on an expansion of Rhetorical Structure Theory (RST)
The framework encompasses discourse relation graphs with tree-breaking, non-projective and concurrent relations, as well as implicit and explicit signals which give explainable rationales to our analyses.
We present and evaluate a freely available corpus of English annotated according to our framework, encompassing 12 spoken and written genres with over 200K tokens.
arXiv Detail & Related papers (2024-03-20T12:52:38Z) - Prototype-based Aleatoric Uncertainty Quantification for Cross-modal
Retrieval [139.21955930418815]
Cross-modal Retrieval methods build similarity relations between vision and language modalities by jointly learning a common representation space.
However, the predictions are often unreliable due to the Aleatoric uncertainty, which is induced by low-quality data, e.g., corrupt images, fast-paced videos, and non-detailed texts.
We propose a novel Prototype-based Aleatoric Uncertainty Quantification (PAU) framework to provide trustworthy predictions by quantifying the uncertainty arisen from the inherent data ambiguity.
arXiv Detail & Related papers (2023-09-29T09:41:19Z) - Testing the Predictions of Surprisal Theory in 11 Languages [77.45204595614]
We investigate the relationship between surprisal and reading times in eleven different languages.
By focusing on a more diverse set of languages, we argue that these results offer the most robust link to-date between information theory and incremental language processing across languages.
arXiv Detail & Related papers (2023-07-07T15:37:50Z) - Token-wise Decomposition of Autoregressive Language Model Hidden States
for Analyzing Model Predictions [9.909170013118775]
This work presents a linear decomposition of final hidden states from autoregressive language models based on each initial input token.
Using the change in next-word probability as a measure of importance, this work first examines which context words make the biggest contribution to language model predictions.
arXiv Detail & Related papers (2023-05-17T23:55:32Z) - Probing for Incremental Parse States in Autoregressive Language Models [9.166953511173903]
Next-word predictions from autoregressive neural language models show remarkable sensitivity to syntax.
This work evaluates the extent to which this behavior arises as a result of a learned ability to maintain implicit representations of incremental syntactic structures.
arXiv Detail & Related papers (2022-11-17T18:15:31Z) - Syntactic Surprisal From Neural Models Predicts, But Underestimates,
Human Processing Difficulty From Syntactic Ambiguities [19.659811811023374]
We propose a method for estimating syntactic predictability from a language model.
We find that treating syntactic predictability independently from lexical predictability indeed results in larger estimates of garden path effects.
Our results support the hypothesis that predictability is not the only factor responsible for the processing cost associated with garden path sentences.
arXiv Detail & Related papers (2022-10-21T18:30:56Z) - Predictive Inference with Feature Conformal Prediction [80.77443423828315]
We propose feature conformal prediction, which extends the scope of conformal prediction to semantic feature spaces.
From a theoretical perspective, we demonstrate that feature conformal prediction provably outperforms regular conformal prediction under mild assumptions.
Our approach could be combined with not only vanilla conformal prediction, but also other adaptive conformal prediction methods.
arXiv Detail & Related papers (2022-10-01T02:57:37Z) - Compositional Generalization Requires Compositional Parsers [69.77216620997305]
We compare sequence-to-sequence models and models guided by compositional principles on the recent COGS corpus.
We show structural generalization is a key measure of compositional generalization and requires models that are aware of complex structure.
arXiv Detail & Related papers (2022-02-24T07:36:35Z) - Instance-Based Neural Dependency Parsing [56.63500180843504]
We develop neural models that possess an interpretable inference process for dependency parsing.
Our models adopt instance-based inference, where dependency edges are extracted and labeled by comparing them to edges in a training set.
arXiv Detail & Related papers (2021-09-28T05:30:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.