Merge-based syntax is mediated by distinct neurocognitive mechanisms: A clustering analysis of comprehension abilities in 84,000 individuals with language deficits across nine languages
- URL: http://arxiv.org/abs/2508.02885v1
- Date: Mon, 04 Aug 2025 20:33:36 GMT
- Title: Merge-based syntax is mediated by distinct neurocognitive mechanisms: A clustering analysis of comprehension abilities in 84,000 individuals with language deficits across nine languages
- Authors: Elliot Murphy, Rohan Venkatesh, Edward Khokhlovich, Andrey Vyshedskiy,
- Abstract summary: Merge is an elementary, indivisible operation that emerged in a single evolutionary step.<n>From a neurocognitive standpoint, different mental objects constructed by Merge may be supported by distinct mechanisms.<n>While a Merge-based syntax may still have emerged suddenly in evolutionary time, different cognitive mechanisms seem to underwrite the processing of various types of Merge-based objects.
- Score: 0.8437187555622164
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: In the modern language sciences, the core computational operation of syntax, 'Merge', is defined as an operation that combines two linguistic units (e.g., 'brown', 'cat') to form a categorized structure ('brown cat', a Noun Phrase). This can then be further combined with additional linguistic units based on this categorial information, respecting non-associativity such that abstract grouping is respected. Some linguists have embraced the view that Merge is an elementary, indivisible operation that emerged in a single evolutionary step. From a neurocognitive standpoint, different mental objects constructed by Merge may be supported by distinct mechanisms: (1) simple command constructions (e.g., "eat apples"); (2) the merging of adjectives and nouns ("red boat"); and (3) the merging of nouns with spatial prepositions ("laptop behind the sofa"). Here, we systematically investigate participants' comprehension of sentences with increasing levels of syntactic complexity. Clustering analyses revealed behavioral evidence for three distinct structural types, which we discuss as potentially emerging at different developmental stages and subject to selective impairment. While a Merge-based syntax may still have emerged suddenly in evolutionary time, responsible for the structured symbolic turn our species took, different cognitive mechanisms seem to underwrite the processing of various types of Merge-based objects.
Related papers
- Objective-Free Local Learning and Emergent Language Structure in Thinking Machines [0.0]
We present a neuro-symbolic framework for generative language modeling based on local, event-driven emergent learning.<n>At its core is a hierarchical Hopfield memory chain acting as a compositional short-term memory and dynamic tokenizer.<n>We demonstrate that briefly activating a new neuron during inference binds distributed multi-scale token features into a symbolic embedding.
arXiv Detail & Related papers (2025-06-29T15:29:13Z) - Agentivit\`a e telicit\`a in GilBERTo: implicazioni cognitive [77.71680953280436]
The goal of this study is to investigate whether a Transformer-based neural language model infers lexical semantics.
The semantic properties considered are telicity (also combined with definiteness) and agentivity.
arXiv Detail & Related papers (2023-07-06T10:52:22Z) - Geometry of Language [0.0]
We present a fresh perspective on language, combining ideas from various sources, but mixed in a new synthesis.
The question is whether we can formulate an elegant formalism, a universal grammar or a mechanism which explains significant aspects of the human faculty of language.
We describe such a mechanism, which differs from existing logical and grammatical approaches by its geometric nature.
arXiv Detail & Related papers (2023-03-09T12:22:28Z) - Composition, Attention, or Both? [8.22379888383833]
We propose a novel architecture called Composition Attention Grammars (CAGs)
We investigate whether composition function and self-attention mechanism can both induce human-like syntactic generalization.
arXiv Detail & Related papers (2022-10-24T05:30:02Z) - Center-Embedding and Constituency in the Brain and a New
Characterization of Context-Free Languages [2.8932261919131017]
We show that constituency and the processing of dependent sentences can be implemented by neurons and synapses.
Surprisingly, the way we implement center embedding points to a new characterization of context-free languages.
arXiv Detail & Related papers (2022-06-27T12:11:03Z) - Unsupervised Learning of Hierarchical Conversation Structure [50.29889385593043]
Goal-oriented conversations often have meaningful sub-dialogue structure, but it can be highly domain-dependent.
This work introduces an unsupervised approach to learning hierarchical conversation structure, including turn and sub-dialogue segment labels.
The decoded structure is shown to be useful in enhancing neural models of language for three conversation-level understanding tasks.
arXiv Detail & Related papers (2022-05-24T17:52:34Z) - Emergence of Machine Language: Towards Symbolic Intelligence with Neural
Networks [73.94290462239061]
We propose to combine symbolism and connectionism principles by using neural networks to derive a discrete representation.
By designing an interactive environment and task, we demonstrated that machines could generate a spontaneous, flexible, and semantic language.
arXiv Detail & Related papers (2022-01-14T14:54:58Z) - Low-Dimensional Structure in the Space of Language Representations is
Reflected in Brain Responses [62.197912623223964]
We show a low-dimensional structure where language models and translation models smoothly interpolate between word embeddings, syntactic and semantic tasks, and future word embeddings.
We find that this representation embedding can predict how well each individual feature space maps to human brain responses to natural language stimuli recorded using fMRI.
This suggests that the embedding captures some part of the brain's natural language representation structure.
arXiv Detail & Related papers (2021-06-09T22:59:12Z) - Compositional Processing Emerges in Neural Networks Solving Math
Problems [100.80518350845668]
Recent progress in artificial neural networks has shown that when large models are trained on enough linguistic data, grammatical structure emerges in their representations.
We extend this work to the domain of mathematical reasoning, where it is possible to formulate precise hypotheses about how meanings should be composed.
Our work shows that neural networks are not only able to infer something about the structured relationships implicit in their training data, but can also deploy this knowledge to guide the composition of individual meanings into composite wholes.
arXiv Detail & Related papers (2021-05-19T07:24:42Z) - Decomposing lexical and compositional syntax and semantics with deep
language models [82.81964713263483]
The activations of language transformers like GPT2 have been shown to linearly map onto brain activity during speech comprehension.
Here, we propose a taxonomy to factorize the high-dimensional activations of language models into four classes: lexical, compositional, syntactic, and semantic representations.
The results highlight two findings. First, compositional representations recruit a more widespread cortical network than lexical ones, and encompass the bilateral temporal, parietal and prefrontal cortices.
arXiv Detail & Related papers (2021-03-02T10:24:05Z) - Seeing Both the Forest and the Trees: Multi-head Attention for Joint
Classification on Different Compositional Levels [15.453888735879525]
In natural languages, words are used in association to construct sentences.
We design a deep neural network architecture that explicitly wires lower and higher linguistic components.
We show that our model, MHAL, learns to simultaneously solve them at different levels of granularity.
arXiv Detail & Related papers (2020-11-01T10:44:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.