Brain2Word: Decoding Brain Activity for Language Generation
- URL: http://arxiv.org/abs/2009.04765v3
- Date: Wed, 11 Nov 2020 08:07:08 GMT
- Title: Brain2Word: Decoding Brain Activity for Language Generation
- Authors: Nicolas Affolter, Beni Egressy, Damian Pascual, Roger Wattenhofer
- Abstract summary: We present a model that can decode fMRI data from unseen subjects.
Our model achieves 5.22% Top-1 and 13.59% Top-5 accuracy in this challenging task.
- Score: 14.24200473508597
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Brain decoding, understood as the process of mapping brain activities to the
stimuli that generated them, has been an active research area in the last
years. In the case of language stimuli, recent studies have shown that it is
possible to decode fMRI scans into an embedding of the word a subject is
reading. However, such word embeddings are designed for natural language
processing tasks rather than for brain decoding. Therefore, they limit our
ability to recover the precise stimulus. In this work, we propose to directly
classify an fMRI scan, mapping it to the corresponding word within a fixed
vocabulary. Unlike existing work, we evaluate on scans from previously unseen
subjects. We argue that this is a more realistic setup and we present a model
that can decode fMRI data from unseen subjects. Our model achieves 5.22% Top-1
and 13.59% Top-5 accuracy in this challenging task, significantly outperforming
all the considered competitive baselines. Furthermore, we use the decoded words
to guide language generation with the GPT-2 model. This way, we advance the
quest for a system that translates brain activities into coherent text.
Related papers
- How Many Bytes Can You Take Out Of Brain-To-Text Decoding? [45.665946951551746]
We propose an information-based evaluation metric for brain-to-text decoders.
We show two methods to augment existing state-of-the-art continuous text decoders.
We conclude that a practical brain-to-text decoder is likely possible given further algorithmic improvements.
arXiv Detail & Related papers (2024-05-22T22:57:04Z) - Language Reconstruction with Brain Predictive Coding from fMRI Data [28.217967547268216]
Theory of predictive coding suggests that human brain naturally engages in continuously predicting future word representations.
textscPredFT achieves current state-of-the-art decoding performance with a maximum BLEU-1 score of $27.8%$.
arXiv Detail & Related papers (2024-05-19T16:06:02Z) - MapGuide: A Simple yet Effective Method to Reconstruct Continuous Language from Brain Activities [28.498085040417813]
We propose a method that guides text reconstruction by comparing them with predicted text embeddings mapped from brain activities.
Our method significantly outperforms the current state-of-the-art model, showing average improvements of 77% and 54% on BLEU and METEOR scores.
Such insight can simplify the task of reconstructing language from brain activities for future work.
arXiv Detail & Related papers (2024-03-26T09:18:59Z) - Language Generation from Brain Recordings [68.97414452707103]
We propose a generative language BCI that utilizes the capacity of a large language model and a semantic brain decoder.
The proposed model can generate coherent language sequences aligned with the semantic content of visual or auditory language stimuli.
Our findings demonstrate the potential and feasibility of employing BCIs in direct language generation.
arXiv Detail & Related papers (2023-11-16T13:37:21Z) - Chat2Brain: A Method for Mapping Open-Ended Semantic Queries to Brain
Activation Maps [59.648646222905235]
We propose a method called Chat2Brain that combines LLMs to basic text-2-image model, known as Text2Brain, to map semantic queries to brain activation maps.
We demonstrate that Chat2Brain can synthesize plausible neural activation patterns for more complex tasks of text queries.
arXiv Detail & Related papers (2023-09-10T13:06:45Z) - Toward a realistic model of speech processing in the brain with
self-supervised learning [67.7130239674153]
Self-supervised algorithms trained on the raw waveform constitute a promising candidate.
We show that Wav2Vec 2.0 learns brain-like representations with as little as 600 hours of unlabelled speech.
arXiv Detail & Related papers (2022-06-03T17:01:46Z) - Neural Language Taskonomy: Which NLP Tasks are the most Predictive of
fMRI Brain Activity? [3.186888145772382]
Several popular Transformer based language models have been found to be successful for text-driven brain encoding.
In this work, we explore transfer learning from representations learned for ten popular natural language processing tasks.
Experiments across all 10 task representations provide the following cognitive insights.
arXiv Detail & Related papers (2022-05-03T10:23:08Z) - Open Vocabulary Electroencephalography-To-Text Decoding and Zero-shot
Sentiment Classification [78.120927891455]
State-of-the-art brain-to-text systems have achieved great success in decoding language directly from brain signals using neural networks.
In this paper, we extend the problem to open vocabulary Electroencephalography(EEG)-To-Text Sequence-To-Sequence decoding and zero-shot sentence sentiment classification on natural reading tasks.
Our model achieves a 40.1% BLEU-1 score on EEG-To-Text decoding and a 55.6% F1 score on zero-shot EEG-based ternary sentiment classification, which significantly outperforms supervised baselines.
arXiv Detail & Related papers (2021-12-05T21:57:22Z) - Model-based analysis of brain activity reveals the hierarchy of language
in 305 subjects [82.81964713263483]
A popular approach to decompose the neural bases of language consists in correlating, across individuals, the brain responses to different stimuli.
Here, we show that a model-based approach can reach equivalent results within subjects exposed to natural stimuli.
arXiv Detail & Related papers (2021-10-12T15:30:21Z) - Low-Dimensional Structure in the Space of Language Representations is
Reflected in Brain Responses [62.197912623223964]
We show a low-dimensional structure where language models and translation models smoothly interpolate between word embeddings, syntactic and semantic tasks, and future word embeddings.
We find that this representation embedding can predict how well each individual feature space maps to human brain responses to natural language stimuli recorded using fMRI.
This suggests that the embedding captures some part of the brain's natural language representation structure.
arXiv Detail & Related papers (2021-06-09T22:59:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.