A unified information-theoretic model of EEG signatures of human
language processing
- URL: http://arxiv.org/abs/2212.08205v1
- Date: Fri, 16 Dec 2022 00:15:45 GMT
- Title: A unified information-theoretic model of EEG signatures of human
language processing
- Authors: Jiaxuan Li and Richard Futrell
- Abstract summary: We advance an information-theoretic model of human language processing in the brain, in which incoming linguistic input is processed at two levels.
We propose that these two kinds of information processing have distinct electroencephalographic signatures, corresponding to the well-documented N400 and P600 components of language-related event-related potentials (ERPs)
Our theory is in principle compatible with traditional cognitive theories assuming a good-enough' interpretation stage, but with precise information-theoretic formulation.
- Score: 7.190747604294439
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We advance an information-theoretic model of human language processing in the
brain, in which incoming linguistic input is processed at two levels, in terms
of a heuristic interpretation and in terms of error correction. We propose that
these two kinds of information processing have distinct electroencephalographic
signatures, corresponding to the well-documented N400 and P600 components of
language-related event-related potentials (ERPs). Formally, we show that the
information content (surprisal) of a word in context can be decomposed into two
quantities: (A) heuristic surprise, which signals processing difficulty of word
given its inferred context, and corresponds with the N400 signal; and (B)
discrepancy signal, which reflects divergence between the true context and the
inferred context, and corresponds to the P600 signal. Both of these quantities
can be estimated using modern NLP techniques. We validate our theory by
successfully simulating ERP patterns elicited by a variety of linguistic
manipulations in previously-reported experimental data from Ryskin et al.
(2021). Our theory is in principle compatible with traditional cognitive
theories assuming a `good-enough' heuristic interpretation stage, but with
precise information-theoretic formulation.
Related papers
- Decomposition of surprisal: Unified computational model of ERP components in language processing [7.760815504640362]
We advance an information-theoretic model of human language processing in the brain in which incoming linguistic input is processed at first shallowly and later with more depth.
We show that the information content (surprisal) of a word in context can be decomposed into two quantities: (A) surprise, which signals shallow processing difficulty for a word, and (B) discrepancy signal, which reflects the discrepancy between shallow and deep interpretations.
arXiv Detail & Related papers (2024-09-10T18:14:02Z) - An information-theoretic model of shallow and deep language comprehension [7.760815504640362]
We use information theory to formulate a model of language comprehension as an optimal trade-off between accuracy and processing depth.
By quantifying the timecourse of language processing as it proceeds from shallow to deep, our model provides a unified framework to explain behavioral and neural signatures of language comprehension.
arXiv Detail & Related papers (2024-05-13T22:29:33Z) - Explaining Text Similarity in Transformer Models [52.571158418102584]
Recent advances in explainable AI have made it possible to mitigate limitations by leveraging improved explanations for Transformers.
We use BiLRP, an extension developed for computing second-order explanations in bilinear similarity models, to investigate which feature interactions drive similarity in NLP models.
Our findings contribute to a deeper understanding of different semantic similarity tasks and models, highlighting how novel explainable AI methods enable in-depth analyses and corpus-level insights.
arXiv Detail & Related papers (2024-05-10T17:11:31Z) - Brain-Driven Representation Learning Based on Diffusion Model [25.375490061512]
Denoising diffusion probabilistic models (DDPMs) are explored in our research as a means to address this issue.
Using DDPMs in conjunction with a conditional autoencoder, our new approach considerably outperforms traditional machine learning algorithms.
Our results highlight the potential of DDPMs as a sophisticated computational method for the analysis of speech-related EEG signals.
arXiv Detail & Related papers (2023-11-14T05:59:58Z) - HyPoradise: An Open Baseline for Generative Speech Recognition with
Large Language Models [81.56455625624041]
We introduce the first open-source benchmark to utilize external large language models (LLMs) for ASR error correction.
The proposed benchmark contains a novel dataset, HyPoradise (HP), encompassing more than 334,000 pairs of N-best hypotheses.
LLMs with reasonable prompt and its generative capability can even correct those tokens that are missing in N-best list.
arXiv Detail & Related papers (2023-09-27T14:44:10Z) - Constructing Word-Context-Coupled Space Aligned with Associative
Knowledge Relations for Interpretable Language Modeling [0.0]
The black-box structure of the deep neural network in pre-trained language models seriously limits the interpretability of the language modeling process.
A Word-Context-Coupled Space (W2CSpace) is proposed by introducing the alignment processing between uninterpretable neural representation and interpretable statistical logic.
Our language model can achieve better performance and highly credible interpretable ability compared to related state-of-the-art methods.
arXiv Detail & Related papers (2023-05-19T09:26:02Z) - A Unified Understanding of Deep NLP Models for Text Classification [88.35418976241057]
We have developed a visual analysis tool, DeepNLPVis, to enable a unified understanding of NLP models for text classification.
The key idea is a mutual information-based measure, which provides quantitative explanations on how each layer of a model maintains the information of input words in a sample.
A multi-level visualization, which consists of a corpus-level, a sample-level, and a word-level visualization, supports the analysis from the overall training set to individual samples.
arXiv Detail & Related papers (2022-06-19T08:55:07Z) - A Latent-Variable Model for Intrinsic Probing [93.62808331764072]
We propose a novel latent-variable formulation for constructing intrinsic probes.
We find empirical evidence that pre-trained representations develop a cross-lingually entangled notion of morphosyntax.
arXiv Detail & Related papers (2022-01-20T15:01:12Z) - CogAlign: Learning to Align Textual Neural Representations to Cognitive
Language Processing Signals [60.921888445317705]
We propose a CogAlign approach to integrate cognitive language processing signals into natural language processing models.
We show that CogAlign achieves significant improvements with multiple cognitive features over state-of-the-art models on public datasets.
arXiv Detail & Related papers (2021-06-10T07:10:25Z) - Data Augmentation for Spoken Language Understanding via Pretrained
Language Models [113.56329266325902]
Training of spoken language understanding (SLU) models often faces the problem of data scarcity.
We put forward a data augmentation method using pretrained language models to boost the variability and accuracy of generated utterances.
arXiv Detail & Related papers (2020-04-29T04:07:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.