Identifying and interpreting non-aligned human conceptual
representations using language modeling
- URL: http://arxiv.org/abs/2403.06204v1
- Date: Sun, 10 Mar 2024 13:02:27 GMT
- Title: Identifying and interpreting non-aligned human conceptual
representations using language modeling
- Authors: Wanqian Bao and Uri Hasson
- Abstract summary: We show that congenital blindness induces conceptual reorganization in both a-modal and sensory-related verbal domains.
We find that blind individuals more strongly associate social and cognitive meanings to verbs related to motion.
For some verbs, representations of blind and sighted are highly similar.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The question of whether people's experience in the world shapes conceptual
representation and lexical semantics is longstanding. Word-association,
feature-listing and similarity rating tasks aim to address this question but
require a subjective interpretation of the latent dimensions identified. In
this study, we introduce a supervised representational-alignment method that
(i) determines whether two groups of individuals share the same basis of a
certain category, and (ii) explains in what respects they differ. In applying
this method, we show that congenital blindness induces conceptual
reorganization in both a-modal and sensory-related verbal domains, and we
identify the associated semantic shifts. We first apply supervised
feature-pruning to a language model (GloVe) to optimize prediction accuracy of
human similarity judgments from word embeddings. Pruning identifies one subset
of retained GloVe features that optimizes prediction of judgments made by
sighted individuals and another subset that optimizes judgments made by blind.
A linear probing analysis then interprets the latent semantics of these
feature-subsets by learning a mapping from the retained GloVe features to 65
interpretable semantic dimensions. We applied this approach to seven semantic
domains, including verbs related to motion, sight, touch, and amodal verbs
related to knowledge acquisition. We find that blind individuals more strongly
associate social and cognitive meanings to verbs related to motion or those
communicating non-speech vocal utterances (e.g., whimper, moan). Conversely,
for amodal verbs, they demonstrate much sparser information. Finally, for some
verbs, representations of blind and sighted are highly similar. The study
presents a formal approach for studying interindividual differences in word
meaning, and the first demonstration of how blindness impacts conceptual
representation of everyday verbs.
Related papers
- Pixel Sentence Representation Learning [67.4775296225521]
In this work, we conceptualize the learning of sentence-level textual semantics as a visual representation learning process.
We employ visually-grounded text perturbation methods like typos and word order shuffling, resonating with human cognitive patterns, and enabling perturbation to be perceived as continuous.
Our approach is further bolstered by large-scale unsupervised topical alignment training and natural language inference supervision.
arXiv Detail & Related papers (2024-02-13T02:46:45Z) - Enhancing Interpretability using Human Similarity Judgements to Prune
Word Embeddings [0.0]
Interpretability methods in NLP aim to provide insights into the semantics underlying specific system architectures.
We present a supervised-learning method that identifies a subset of model features that strongly improve prediction of human similarity judgments.
We show this method keeps only 20-40% of the original embeddings, for 8 independent semantic domains.
We then present two approaches for interpreting the semantics of the retained features.
arXiv Detail & Related papers (2023-10-16T10:38:49Z) - Agentivit\`a e telicit\`a in GilBERTo: implicazioni cognitive [77.71680953280436]
The goal of this study is to investigate whether a Transformer-based neural language model infers lexical semantics.
The semantic properties considered are telicity (also combined with definiteness) and agentivity.
arXiv Detail & Related papers (2023-07-06T10:52:22Z) - Neighboring Words Affect Human Interpretation of Saliency Explanations [65.29015910991261]
Word-level saliency explanations are often used to communicate feature-attribution in text-based models.
Recent studies found that superficial factors such as word length can distort human interpretation of the communicated saliency scores.
We investigate how the marking of a word's neighboring words affect the explainee's perception of the word's importance in the context of a saliency explanation.
arXiv Detail & Related papers (2023-05-04T09:50:25Z) - Visual Superordinate Abstraction for Robust Concept Learning [80.15940996821541]
Concept learning constructs visual representations that are connected to linguistic semantics.
We ascribe the bottleneck to a failure of exploring the intrinsic semantic hierarchy of visual concepts.
We propose a visual superordinate abstraction framework for explicitly modeling semantic-aware visual subspaces.
arXiv Detail & Related papers (2022-05-28T14:27:38Z) - Learnable Visual Words for Interpretable Image Recognition [70.85686267987744]
We propose the Learnable Visual Words (LVW) to interpret the model prediction behaviors with two novel modules.
The semantic visual words learning relaxes the category-specific constraint, enabling the general visual words shared across different categories.
Our experiments on six visual benchmarks demonstrate the superior effectiveness of our proposed LVW in both accuracy and model interpretation.
arXiv Detail & Related papers (2022-05-22T03:24:45Z) - Building a visual semantics aware object hierarchy [0.0]
We propose a novel unsupervised method to build visual semantics aware object hierarchy.
Our intuition in this paper comes from real-world knowledge representation where concepts are hierarchically organized.
The evaluation consists of two parts, firstly we apply the constructed hierarchy on the object recognition task and then we compare our visual hierarchy and existing lexical hierarchies to show the validity of our method.
arXiv Detail & Related papers (2022-02-26T00:10:21Z) - Unsupervised Multimodal Word Discovery based on Double Articulation
Analysis with Co-occurrence cues [7.332652485849632]
Human infants acquire their verbal lexicon with minimal prior knowledge of language.
This study proposes a novel fully unsupervised learning method for discovering speech units.
The proposed method can acquire words and phonemes from speech signals using unsupervised learning.
arXiv Detail & Related papers (2022-01-18T07:31:59Z) - Understanding Synonymous Referring Expressions via Contrastive Features [105.36814858748285]
We develop an end-to-end trainable framework to learn contrastive features on the image and object instance levels.
We conduct extensive experiments to evaluate the proposed algorithm on several benchmark datasets.
arXiv Detail & Related papers (2021-04-20T17:56:24Z) - Hearings and mishearings: decrypting the spoken word [0.0]
We propose a model of the speech perception of individual words in the presence of mishearings.
We show for instance that speech perception is easy when the word length is less than a threshold, to be identified with a static transition.
We extend this to the dynamics of word recognition, proposing an intuitive approach highlighting the distinction between individual, isolated mishearings and clusters of contiguous mishearings.
arXiv Detail & Related papers (2020-09-01T13:58:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.