A Sentence is Worth a Thousand Pictures: Can Large Language Models
Understand Human Language?
- URL: http://arxiv.org/abs/2308.00109v1
- Date: Wed, 26 Jul 2023 18:58:53 GMT
- Title: A Sentence is Worth a Thousand Pictures: Can Large Language Models
Understand Human Language?
- Authors: Gary Marcus, Evelina Leivada, Elliot Murphy
- Abstract summary: We analyze the contribution of large language models as theoretically informative representations of a target system vs. atheoretical powerful mechanistic tools.
We identify the key abilities that are still missing from the current state of development and exploitation of these models.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Artificial Intelligence applications show great potential for
language-related tasks that rely on next-word prediction. The current
generation of large language models have been linked to claims about human-like
linguistic performance and their applications are hailed both as a key step
towards Artificial General Intelligence and as major advance in understanding
the cognitive, and even neural basis of human language. We analyze the
contribution of large language models as theoretically informative
representations of a target system vs. atheoretical powerful mechanistic tools,
and we identify the key abilities that are still missing from the current state
of development and exploitation of these models.
Related papers
- Self Generated Wargame AI: Double Layer Agent Task Planning Based on
Large Language Model [0.6562256987706128]
This paper innovatively applies the large language model to the field of intelligent decision-making.
It proposes a two-layer agent task planning, issues and executes decision commands through the interaction of natural language.
It is found that the intelligent decision-making ability of the large language model is significantly stronger than the commonly used reinforcement learning AI and rule AI.
arXiv Detail & Related papers (2023-12-02T09:45:45Z) - Formal Aspects of Language Modeling [74.16212987886013]
Large language models have become one of the most commonly deployed NLP inventions.
These notes are the accompaniment to the theoretical portion of the ETH Z"urich course on large language models.
arXiv Detail & Related papers (2023-11-07T20:21:42Z) - Large Language Models for Scientific Synthesis, Inference and
Explanation [56.41963802804953]
We show how large language models can perform scientific synthesis, inference, and explanation.
We show that the large language model can augment this "knowledge" by synthesizing from the scientific literature.
This approach has the further advantage that the large language model can explain the machine learning system's predictions.
arXiv Detail & Related papers (2023-10-12T02:17:59Z) - Diffusion Language Models Can Perform Many Tasks with Scaling and
Instruction-Finetuning [56.03057119008865]
We show that scaling diffusion language models can effectively make them strong language learners.
We build competent diffusion language models at scale by first acquiring knowledge from massive data.
Experiments show that scaling diffusion language models consistently improves performance across downstream language tasks.
arXiv Detail & Related papers (2023-08-23T16:01:12Z) - A Survey of Large Language Models [81.06947636926638]
Language modeling has been widely studied for language understanding and generation in the past two decades.
Recently, pre-trained language models (PLMs) have been proposed by pre-training Transformer models over large-scale corpora.
To discriminate the difference in parameter scale, the research community has coined the term large language models (LLM) for the PLMs of significant size.
arXiv Detail & Related papers (2023-03-31T17:28:46Z) - Beyond the limitations of any imaginable mechanism: large language
models and psycholinguistics [0.0]
Large language models provide a model for language.
They are useful as a practical tool, as an illustrative comparative, and philosophical, as a basis for recasting the relationship between language and thought.
arXiv Detail & Related papers (2023-02-28T20:49:38Z) - What Artificial Neural Networks Can Tell Us About Human Language
Acquisition [47.761188531404066]
Rapid progress in machine learning for natural language processing has the potential to transform debates about how humans learn language.
To increase the relevance of learnability results from computational models, we need to train model learners without significant advantages over humans.
arXiv Detail & Related papers (2022-08-17T00:12:37Z) - Language Models are not Models of Language [0.0]
Transfer learning has enabled large deep learning neural networks trained on the language modeling task to vastly improve performance.
We argue that the term language model is misleading because deep learning models are not theoretical models of language.
arXiv Detail & Related papers (2021-12-13T22:39:46Z) - Towards Zero-shot Language Modeling [90.80124496312274]
We construct a neural model that is inductively biased towards learning human languages.
We infer this distribution from a sample of typologically diverse training languages.
We harness additional language-specific side information as distant supervision for held-out languages.
arXiv Detail & Related papers (2021-08-06T23:49:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.