Model-based analysis of brain activity reveals the hierarchy of language
in 305 subjects
- URL: http://arxiv.org/abs/2110.06078v1
- Date: Tue, 12 Oct 2021 15:30:21 GMT
- Title: Model-based analysis of brain activity reveals the hierarchy of language
in 305 subjects
- Authors: Charlotte Caucheteux, Alexandre Gramfort, Jean-R\'emi King
- Abstract summary: A popular approach to decompose the neural bases of language consists in correlating, across individuals, the brain responses to different stimuli.
Here, we show that a model-based approach can reach equivalent results within subjects exposed to natural stimuli.
- Score: 82.81964713263483
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: A popular approach to decompose the neural bases of language consists in
correlating, across individuals, the brain responses to different stimuli (e.g.
regular speech versus scrambled words, sentences, or paragraphs). Although
successful, this `model-free' approach necessitates the acquisition of a large
and costly set of neuroimaging data. Here, we show that a model-based approach
can reach equivalent results within subjects exposed to natural stimuli. We
capitalize on the recently-discovered similarities between deep language models
and the human brain to compute the mapping between i) the brain responses to
regular speech and ii) the activations of deep language models elicited by
modified stimuli (e.g. scrambled words, sentences, or paragraphs). Our
model-based approach successfully replicates the seminal study of Lerner et al.
(2011), which revealed the hierarchy of language areas by comparing the
functional-magnetic resonance imaging (fMRI) of seven subjects listening to
7min of both regular and scrambled narratives. We further extend and precise
these results to the brain signals of 305 individuals listening to 4.1 hours of
narrated stories. Overall, this study paves the way for efficient and flexible
analyses of the brain bases of language.
Related papers
- Brain-Like Language Processing via a Shallow Untrained Multihead Attention Network [16.317199232071232]
Large Language Models (LLMs) have been shown to be effective models of the human language system.
In this work, we investigate the key architectural components driving the surprising alignment of untrained models.
arXiv Detail & Related papers (2024-06-21T12:54:03Z) - fMRI predictors based on language models of increasing complexity recover brain left lateralization [4.1618731507412505]
We show that the left-right difference in brain correlation follows a scaling law with the number of parameters.
This finding reconciles computational analyses of brain activity using large language models with the classic observation from aphasic patients showing left hemisphere dominance for language.
arXiv Detail & Related papers (2024-05-28T09:24:52Z) - Navigating Brain Language Representations: A Comparative Analysis of Neural Language Models and Psychologically Plausible Models [29.50162863143141]
We compare encoding performance of various neural language models and psychologically plausible models.
Surprisingly, our findings revealed that psychologically plausible models outperformed neural language models across diverse contexts.
arXiv Detail & Related papers (2024-04-30T08:48:07Z) - Causal Graph in Language Model Rediscovers Cortical Hierarchy in Human
Narrative Processing [0.0]
Previous studies have demonstrated that the features of language models can be mapped to fMRI brain activity.
This raises the question: is there a commonality between information processing in language models and the human brain?
To estimate information flow patterns in a language model, we examined the causal relationships between different layers.
arXiv Detail & Related papers (2023-11-17T10:09:12Z) - Language Generation from Brain Recordings [68.97414452707103]
We propose a generative language BCI that utilizes the capacity of a large language model and a semantic brain decoder.
The proposed model can generate coherent language sequences aligned with the semantic content of visual or auditory language stimuli.
Our findings demonstrate the potential and feasibility of employing BCIs in direct language generation.
arXiv Detail & Related papers (2023-11-16T13:37:21Z) - Decoding speech perception from non-invasive brain recordings [48.46819575538446]
We introduce a model trained with contrastive-learning to decode self-supervised representations of perceived speech from non-invasive recordings.
Our model can identify, from 3 seconds of MEG signals, the corresponding speech segment with up to 41% accuracy out of more than 1,000 distinct possibilities.
arXiv Detail & Related papers (2022-08-25T10:01:43Z) - Neural Language Models are not Born Equal to Fit Brain Data, but
Training Helps [75.84770193489639]
We examine the impact of test loss, training corpus and model architecture on the prediction of functional Magnetic Resonance Imaging timecourses of participants listening to an audiobook.
We find that untrained versions of each model already explain significant amount of signal in the brain by capturing similarity in brain responses across identical words.
We suggest good practices for future studies aiming at explaining the human language system using neural language models.
arXiv Detail & Related papers (2022-07-07T15:37:17Z) - Toward a realistic model of speech processing in the brain with
self-supervised learning [67.7130239674153]
Self-supervised algorithms trained on the raw waveform constitute a promising candidate.
We show that Wav2Vec 2.0 learns brain-like representations with as little as 600 hours of unlabelled speech.
arXiv Detail & Related papers (2022-06-03T17:01:46Z) - Long-range and hierarchical language predictions in brains and
algorithms [82.81964713263483]
We show that while deep language algorithms are optimized to predict adjacent words, the human brain would be tuned to make long-range and hierarchical predictions.
This study strengthens predictive coding theory and suggests a critical role of long-range and hierarchical predictions in natural language processing.
arXiv Detail & Related papers (2021-11-28T20:26:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.