Conceptual structure coheres in human cognition but not in large
language models
- URL: http://arxiv.org/abs/2304.02754v2
- Date: Fri, 10 Nov 2023 17:42:31 GMT
- Title: Conceptual structure coheres in human cognition but not in large
language models
- Authors: Siddharth Suresh, Kushin Mukherjee, Xizheng Yu, Wei-Chun Huang, Lisa
Padua, and Timothy T Rogers
- Abstract summary: We show that conceptual structure is robust to differences in culture, language, and method of estimation.
Results highlight an important difference between contemporary large language models and human cognition.
- Score: 7.405352374343134
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Neural network models of language have long been used as a tool for
developing hypotheses about conceptual representation in the mind and brain.
For many years, such use involved extracting vector-space representations of
words and using distances among these to predict or understand human behavior
in various semantic tasks. Contemporary large language models (LLMs), however,
make it possible to interrogate the latent structure of conceptual
representations using experimental methods nearly identical to those commonly
used with human participants. The current work utilizes three common techniques
borrowed from cognitive psychology to estimate and compare the structure of
concepts in humans and a suite of LLMs. In humans, we show that conceptual
structure is robust to differences in culture, language, and method of
estimation. Structures estimated from LLM behavior, while individually fairly
consistent with those estimated from human behavior, vary much more depending
upon the particular task used to generate responses--across tasks, estimates of
conceptual structure from the very same model cohere less with one another than
do human structure estimates. These results highlight an important difference
between contemporary LLMs and human cognition, with implications for
understanding some fundamental limitations of contemporary machine language.
Related papers
- HLB: Benchmarking LLMs' Humanlikeness in Language Use [2.438748974410787]
We present a comprehensive humanlikeness benchmark (HLB) evaluating 20 large language models (LLMs)
We collected responses from over 2,000 human participants and compared them to outputs from the LLMs in these experiments.
Our results reveal fine-grained differences in how well LLMs replicate human responses across various linguistic levels.
arXiv Detail & Related papers (2024-09-24T09:02:28Z) - Human-like object concept representations emerge naturally in multimodal large language models [24.003766123531545]
We combined behavioral and neuroimaging analysis methods to uncover how the object concept representations in Large Language Models correlate with those of humans.
The resulting 66-dimensional embeddings were found to be highly stable and predictive, and exhibited semantic clustering akin to human mental representations.
This study advances our understanding of machine intelligence and informs the development of more human-like artificial cognitive systems.
arXiv Detail & Related papers (2024-07-01T08:17:19Z) - ConcEPT: Concept-Enhanced Pre-Training for Language Models [57.778895980999124]
ConcEPT aims to infuse conceptual knowledge into pre-trained language models.
It exploits external entity concept prediction to predict the concepts of entities mentioned in the pre-training contexts.
Results of experiments show that ConcEPT gains improved conceptual knowledge with concept-enhanced pre-training.
arXiv Detail & Related papers (2024-01-11T05:05:01Z) - Interpreting Pretrained Language Models via Concept Bottlenecks [55.47515772358389]
Pretrained language models (PLMs) have made significant strides in various natural language processing tasks.
The lack of interpretability due to their black-box'' nature poses challenges for responsible implementation.
We propose a novel approach to interpreting PLMs by employing high-level, meaningful concepts that are easily understandable for humans.
arXiv Detail & Related papers (2023-11-08T20:41:18Z) - Semantic Feature Verification in FLAN-T5 [0.0]
We show that machine-verified norms capture aspects of conceptual structure beyond what is expressed in human norms alone.
The results suggest that LLMs can greatly enhance traditional methods of semantic feature norm verification.
arXiv Detail & Related papers (2023-04-12T03:37:57Z) - A Minimalist Dataset for Systematic Generalization of Perception,
Syntax, and Semantics [131.93113552146195]
We present a new dataset, Handwritten arithmetic with INTegers (HINT), to examine machines' capability of learning generalizable concepts.
In HINT, machines are tasked with learning how concepts are perceived from raw signals such as images.
We undertake extensive experiments with various sequence-to-sequence models, including RNNs, Transformers, and GPT-3.
arXiv Detail & Related papers (2021-03-02T01:32:54Z) - Bongard-LOGO: A New Benchmark for Human-Level Concept Learning and
Reasoning [78.13740873213223]
Bongard problems (BPs) were introduced as an inspirational challenge for visual cognition in intelligent systems.
We propose a new benchmark Bongard-LOGO for human-level concept learning and reasoning.
arXiv Detail & Related papers (2020-10-02T03:19:46Z) - Concept Learners for Few-Shot Learning [76.08585517480807]
We propose COMET, a meta-learning method that improves generalization ability by learning to learn along human-interpretable concept dimensions.
We evaluate our model on few-shot tasks from diverse domains, including fine-grained image classification, document categorization and cell type annotation.
arXiv Detail & Related papers (2020-07-14T22:04:17Z) - Mechanisms for Handling Nested Dependencies in Neural-Network Language
Models and Humans [75.15855405318855]
We studied whether a modern artificial neural network trained with "deep learning" methods mimics a central aspect of human sentence processing.
Although the network was solely trained to predict the next word in a large corpus, analysis showed the emergence of specialized units that successfully handled local and long-distance syntactic agreement.
We tested the model's predictions in a behavioral experiment where humans detected violations in number agreement in sentences with systematic variations in the singular/plural status of multiple nouns.
arXiv Detail & Related papers (2020-06-19T12:00:05Z) - The Fluidity of Concept Representations in Human Brain Signals [0.0]
We analyze the discriminability of concrete and abstract concepts in fMRI data.
We argue that fluid concept representations lead to more realistic models of human language processing.
arXiv Detail & Related papers (2020-02-20T17:31:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.