Why Linguistics Will Thrive in the 21st Century: A Reply to Piantadosi
(2023)
- URL: http://arxiv.org/abs/2308.03228v1
- Date: Sun, 6 Aug 2023 23:41:14 GMT
- Title: Why Linguistics Will Thrive in the 21st Century: A Reply to Piantadosi
(2023)
- Authors: Jordan Kodner, Sarah Payne, Jeffrey Heinz
- Abstract summary: We present a critical assessment of Piantadosi's claim that "Modern language models refute Chomsky's approach to language"
Despite the impressive performance and utility of large language models, humans achieve their capacity for language after exposure to several orders of magnitude less data.
We conclude that generative linguistics as a scientific discipline will remain indispensable throughout the 21st century and beyond.
- Score: 5.2424255020469595
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We present a critical assessment of Piantadosi's (2023) claim that "Modern
language models refute Chomsky's approach to language," focusing on four main
points. First, despite the impressive performance and utility of large language
models (LLMs), humans achieve their capacity for language after exposure to
several orders of magnitude less data. The fact that young children become
competent, fluent speakers of their native languages with relatively little
exposure to them is the central mystery of language learning to which Chomsky
initially drew attention, and LLMs currently show little promise of solving
this mystery. Second, what can the artificial reveal about the natural? Put
simply, the implications of LLMs for our understanding of the cognitive
structures and mechanisms underlying language and its acquisition are like the
implications of airplanes for understanding how birds fly. Third, LLMs cannot
constitute scientific theories of language for several reasons, not least of
which is that scientific theories must provide interpretable explanations, not
just predictions. This leads to our final point: to even determine whether the
linguistic and cognitive capabilities of LLMs rival those of humans requires
explicating what humans' capacities actually are. In other words, it requires a
separate theory of language and cognition; generative linguistics provides
precisely such a theory. As such, we conclude that generative linguistics as a
scientific discipline will remain indispensable throughout the 21st century and
beyond.
Related papers
- XToM: Exploring the Multilingual Theory of Mind for Large Language Models [57.9821865189077]
Existing evaluations of Theory of Mind in LLMs are largely limited to English.<n>We present XToM, a rigorously validated multilingual benchmark that evaluates ToM across five languages.<n>Our findings expose limitations in LLMs' ability to replicate human-like mentalizing across linguistic contexts.
arXiv Detail & Related papers (2025-06-03T05:23:25Z) - When Less Language is More: Language-Reasoning Disentanglement Makes LLMs Better Multilingual Reasoners [111.50503126693444]
We show that language-specific ablation consistently boosts multilingual reasoning performance.<n>Compared to post-training, our training-free ablation achieves comparable or superior results with minimal computational overhead.
arXiv Detail & Related papers (2025-05-21T08:35:05Z) - On the Thinking-Language Modeling Gap in Large Language Models [68.83670974539108]
We show that there is a significant gap between the modeling of languages and thoughts.<n>We propose a new prompt technique termed Language-of-Thoughts (LoT) to demonstrate and alleviate this gap.
arXiv Detail & Related papers (2025-05-19T09:31:52Z) - Can Language Models Learn Typologically Implausible Languages? [62.823015163987996]
Grammatical features across human languages show intriguing correlations often attributed to learning biases in humans.
We discuss how language models (LMs) allow us to better determine the role of domain-general learning biases in language universals.
We test LMs on an array of highly naturalistic but counterfactual versions of the English (head-initial) and Japanese (head-final) languages.
arXiv Detail & Related papers (2025-02-17T20:40:01Z) - How Linguistics Learned to Stop Worrying and Love the Language Models [17.413438037432414]
We argue that the success of LMs obviates the need for studying linguistic theory and structure.
They force us to rethink arguments about learning and are informative for major questions in linguistic theory.
We offer an optimistic take on the relationship between language models and linguistics.
arXiv Detail & Related papers (2025-01-28T16:13:19Z) - How Do Multilingual Models Remember? Investigating Multilingual Factual Recall Mechanisms [50.13632788453612]
Large Language Models (LLMs) store and retrieve vast amounts of factual knowledge acquired during pre-training.
The question of how these processes generalize to other languages and multilingual LLMs remains unexplored.
We examine when language plays a role in the recall process, uncovering evidence of language-independent and language-dependent mechanisms.
arXiv Detail & Related papers (2024-10-18T11:39:34Z) - Large Models of What? Mistaking Engineering Achievements for Human Linguistic Agency [0.11510009152620666]
We argue that claims regarding linguistic capabilities of Large Language Models (LLMs) are based on at least two unfounded assumptions.
Language completeness assumes that a distinct and complete thing such as a natural language' exists.
The assumption of data completeness relies on the belief that a language can be quantified and wholly captured by data.
arXiv Detail & Related papers (2024-07-11T18:06:01Z) - Language-Specific Neurons: The Key to Multilingual Capabilities in Large Language Models [117.20416338476856]
Large language models (LLMs) demonstrate remarkable multilingual capabilities without being pre-trained on specially curated multilingual parallel corpora.
We propose a novel detection method, language activation probability entropy (LAPE), to identify language-specific neurons within LLMs.
Our findings indicate that LLMs' proficiency in processing a particular language is predominantly due to a small subset of neurons.
arXiv Detail & Related papers (2024-02-26T09:36:05Z) - Language Models Don't Learn the Physical Manifestation of Language [0.3529736140137004]
We argue that language-only models don't learn the physical manifestation of language.
We present an empirical investigation of visual-auditory properties of language through a series of tasks, termed H-Test.
arXiv Detail & Related papers (2024-02-17T17:52:24Z) - Meaning and understanding in large language models [0.0]
Recent developments in the generative large language models (LLMs) of artificial intelligence have led to the belief that traditional philosophical assumptions about machine understanding of language need to be revised.
This article critically evaluates the prevailing tendency to regard machine language performance as mere syntactic manipulation and the simulation of understanding, which is only partial and very shallow, without sufficient referential grounding in the world.
arXiv Detail & Related papers (2023-10-26T14:06:14Z) - Democratizing Reasoning Ability: Tailored Learning from Large Language
Model [97.4921006089966]
We propose a tailored learning approach to distill such reasoning ability to smaller LMs.
We exploit the potential of LLM as a reasoning teacher by building an interactive multi-round learning paradigm.
To exploit the reasoning potential of the smaller LM, we propose self-reflection learning to motivate the student to learn from self-made mistakes.
arXiv Detail & Related papers (2023-10-20T07:50:10Z) - The Quo Vadis of the Relationship between Language and Large Language
Models [3.10770247120758]
Large Language Models (LLMs) have come to encourage the adoption of LLMs as scientific models of language.
We identify the most important theoretical and empirical risks brought about by the adoption of scientific models that lack transparency.
We conclude that, at their current stage of development, LLMs hardly offer any explanations for language.
arXiv Detail & Related papers (2023-10-17T10:54:24Z) - Large Language Models are In-Context Semantic Reasoners rather than
Symbolic Reasoners [75.85554779782048]
Large Language Models (LLMs) have excited the natural language and machine learning community over recent years.
Despite of numerous successful applications, the underlying mechanism of such in-context capabilities still remains unclear.
In this work, we hypothesize that the learned textitsemantics of language tokens do the most heavy lifting during the reasoning process.
arXiv Detail & Related papers (2023-05-24T07:33:34Z) - Verbal behavior without syntactic structures: beyond Skinner and Chomsky [0.0]
We must rediscover the extent to which language is like any other human behavior.
Recent psychological, computational, neurobiological, and evolutionary insights into the shaping and structure of behavior may point us toward a new, viable account of language.
arXiv Detail & Related papers (2023-03-11T00:01:21Z) - Understanding Natural Language Understanding Systems. A Critical
Analysis [91.81211519327161]
The development of machines that guillemotlefttalk like usguillemotright, also known as Natural Language Understanding (NLU) systems, is the Holy Grail of Artificial Intelligence (AI)
But never has the trust that we can build guillemotlefttalking machinesguillemotright been stronger than the one engendered by the last generation of NLU systems.
Are we at the dawn of a new era, in which the Grail is finally closer to us?
arXiv Detail & Related papers (2023-03-01T08:32:55Z) - Shortcut Learning of Large Language Models in Natural Language
Understanding [119.45683008451698]
Large language models (LLMs) have achieved state-of-the-art performance on a series of natural language understanding tasks.
They might rely on dataset bias and artifacts as shortcuts for prediction.
This has significantly affected their generalizability and adversarial robustness.
arXiv Detail & Related papers (2022-08-25T03:51:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.