Still "Talking About Large Language Models": Some Clarifications
- URL: http://arxiv.org/abs/2412.10291v1
- Date: Fri, 13 Dec 2024 17:21:29 GMT
- Title: Still "Talking About Large Language Models": Some Clarifications
- Authors: Murray Shanahan,
- Abstract summary: My paper "Talking About Large Language Models" has more than once been interpreted as advocating a reductionist stance towards large language models.
This short note situates the paper in the context of a larger philosophical project that is concerned with the (mis)use of words rather than metaphysics.
- Score: 13.672268920902187
- License:
- Abstract: My paper "Talking About Large Language Models" has more than once been interpreted as advocating a reductionist stance towards large language models. But the paper was not intended that way, and I do not endorse such positions. This short note situates the paper in the context of a larger philosophical project that is concerned with the (mis)use of words rather than metaphysics, in the spirit of Wittgenstein's later writing.
Related papers
- Large Language Models are Biased Because They Are Large Language Models [7.156628542829618]
This paper seeks to convince the reader that harmful biases are an inevitable consequence arising from the design of large language models.
To the extent that this is true, it suggests that the problem of harmful bias cannot be properly addressed without a serious reconsideration of AI driven by LLMs.
arXiv Detail & Related papers (2024-06-19T01:08:03Z) - Modelling Language [0.0]
This paper argues that large language models have a valuable scientific role to play in serving as scientific models of a language.
It draws upon recent work in philosophy of science to show how large language models could serve as scientific models.
arXiv Detail & Related papers (2024-04-15T08:40:01Z) - mALBERT: Is a Compact Multilingual BERT Model Still Worth It? [5.2116647104135305]
We propose to focus on smaller models, such as compact models like ALBERT, which are more virtuous than these PLMs.
PLMs enable huge breakthroughs in Natural Language Processing tasks, such as Spoken and Natural LanguageUnderstanding, classification, Question-Answering tasks.
Considering these facts, wepropose the free release of the first version of a multilingual compact ALBERT model, pre-trained using Wikipediadata.
arXiv Detail & Related papers (2024-03-27T08:25:28Z) - Formal Aspects of Language Modeling [74.16212987886013]
Large language models have become one of the most commonly deployed NLP inventions.
These notes are the accompaniment to the theoretical portion of the ETH Z"urich course on large language models.
arXiv Detail & Related papers (2023-11-07T20:21:42Z) - Evaluating Large Language Models on Controlled Generation Tasks [92.64781370921486]
We present an extensive analysis of various benchmarks including a sentence planning benchmark with different granularities.
After comparing large language models against state-of-the-start finetuned smaller models, we present a spectrum showing large language models falling behind, are comparable, or exceed the ability of smaller models.
arXiv Detail & Related papers (2023-10-23T03:48:24Z) - Symbolic and Language Agnostic Large Language Models [0.0]
We argue that the relative success of large language models (LLMs) is not a reflection on the symbolic vs. subsymbolic debate.
We suggest here is employing the successful bottom-up strategy in a symbolic setting, producing symbolic, language agnostic and ontologically grounded large language models.
arXiv Detail & Related papers (2023-08-27T20:24:33Z) - Still no evidence for an effect of the proportion of non-native speakers
on language complexity -- A response to Kauhanen, Einhaus & Walkden (2023) [91.3755431537592]
Kauhanen, Einhaus & Walkden criticise results presented in one of my papers.
I show in this rejoinder that both points of criticism are explicitly mentioned and analysed in my paper.
arXiv Detail & Related papers (2023-04-29T09:58:32Z) - Beyond the limitations of any imaginable mechanism: large language
models and psycholinguistics [0.0]
Large language models provide a model for language.
They are useful as a practical tool, as an illustrative comparative, and philosophical, as a basis for recasting the relationship between language and thought.
arXiv Detail & Related papers (2023-02-28T20:49:38Z) - Emergent Abilities of Large Language Models [172.08007363384218]
We consider an ability to be emergent if it is not present in smaller models but is present in larger models.
The existence of such emergence implies that additional scaling could further expand the range of capabilities of language models.
arXiv Detail & Related papers (2022-06-15T17:32:01Z) - Testing the Ability of Language Models to Interpret Figurative Language [69.59943454934799]
Figurative and metaphorical language are commonplace in discourse.
It remains an open question to what extent modern language models can interpret nonliteral phrases.
We introduce Fig-QA, a Winograd-style nonliteral language understanding task.
arXiv Detail & Related papers (2022-04-26T23:42:22Z) - Chain of Thought Prompting Elicits Reasoning in Large Language Models [56.811278668446825]
This paper explores the ability of language models to generate a coherent chain of thought.
Experiments show that inducing a chain of thought via prompting can enable sufficiently large language models to better perform reasoning tasks.
arXiv Detail & Related papers (2022-01-28T02:33:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.