An Evoked Potential-Guided Deep Learning Brain Representation For Visual
Classification
- URL: http://arxiv.org/abs/2006.15357v1
- Date: Sat, 27 Jun 2020 12:46:31 GMT
- Title: An Evoked Potential-Guided Deep Learning Brain Representation For Visual
Classification
- Authors: Xianglin Zheng, Zehong Cao, Quan Bai
- Abstract summary: We propose a deep learning framework guided by the visual evoked potentials, called the Event-Related Potential (ERP)-Long short-term memory (LSTM) framework.
Our results showed that our proposed ERP-LSTM framework could achieve classification accuracies of 66.81% and 27.08% for categories (6 classes) and exemplars (72 classes)
Our findings suggested that decoding visual evoked potentials from EEG signals is an effective strategy to learn discriminative brain representations for visual classification.
- Score: 19.587477797948683
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The new perspective in visual classification aims to decode the feature
representation of visual objects from human brain activities. Recording
electroencephalogram (EEG) from the brain cortex has been seen as a prevalent
approach to understand the cognition process of an image classification task.
In this study, we proposed a deep learning framework guided by the visual
evoked potentials, called the Event-Related Potential (ERP)-Long short-term
memory (LSTM) framework, extracted by EEG signals for visual classification. In
specific, we first extracted the ERP sequences from multiple EEG channels to
response image stimuli-related information. Then, we trained an LSTM network to
learn the feature representation space of visual objects for classification. In
the experiment, 10 subjects were recorded by over 50,000 EEG trials from an
image dataset with 6 categories, including a total of 72 exemplars. Our results
showed that our proposed ERP-LSTM framework could achieve classification
accuracies of cross-subject of 66.81% and 27.08% for categories (6 classes) and
exemplars (72 classes), respectively. Our results outperformed that of using
the existing visual classification frameworks, by improving classification
accuracies in the range of 12.62% - 53.99%. Our findings suggested that
decoding visual evoked potentials from EEG signals is an effective strategy to
learn discriminative brain representations for visual classification.
Related papers
- Visual Neural Decoding via Improved Visual-EEG Semantic Consistency [3.4061238650474657]
Methods that directly map EEG features to the CLIP embedding space may introduce mapping bias and cause semantic inconsistency.
We propose a Visual-EEG Semantic Decouple Framework that explicitly extracts the semantic-related features of these two modalities to facilitate optimal alignment.
Our method achieves state-of-the-art results in zero-shot neural decoding tasks.
arXiv Detail & Related papers (2024-08-13T10:16:10Z) - Knowledge-Guided Prompt Learning for Lifespan Brain MR Image Segmentation [53.70131202548981]
We present a two-step segmentation framework employing Knowledge-Guided Prompt Learning (KGPL) for brain MRI.
Specifically, we first pre-train segmentation models on large-scale datasets with sub-optimal labels.
The introduction of knowledge-wise prompts captures semantic relationships between anatomical variability and biological processes.
arXiv Detail & Related papers (2024-07-31T04:32:43Z) - Mind's Eye: Image Recognition by EEG via Multimodal Similarity-Keeping Contrastive Learning [2.087148326341881]
This paper introduces a MUltimodal Similarity-keeping contrastivE learning framework for zero-shot EEG-based image classification.
We develop a series of multivariate time-series encoders tailored for EEG signals and assess the efficacy of regularized contrastive EEG-Image pretraining.
Our method achieves state-of-the-art performance, with a top-1 accuracy of 19.3% and a top-5 accuracy of 48.8% in 200-way zero-shot image classification.
arXiv Detail & Related papers (2024-06-05T16:42:23Z) - Personality Trait Recognition using ECG Spectrograms and Deep Learning [6.6157730528755065]
This paper presents an innovative approach to recognizing personality traits using deep learning (DL) methods applied to electrocardiogram (ECG) signals.
Within the framework of detecting the big five personality traits model encompassing extra-version, neuroticism, agreeableness, conscientiousness, and openness, the research explores the potential of ECG-derived spectrograms as informative features.
arXiv Detail & Related papers (2024-02-06T19:09:44Z) - MLIP: Enhancing Medical Visual Representation with Divergence Encoder
and Knowledge-guided Contrastive Learning [48.97640824497327]
We propose a novel framework leveraging domain-specific medical knowledge as guiding signals to integrate language information into the visual domain through image-text contrastive learning.
Our model includes global contrastive learning with our designed divergence encoder, local token-knowledge-patch alignment contrastive learning, and knowledge-guided category-level contrastive learning with expert knowledge.
Notably, MLIP surpasses state-of-the-art methods even with limited annotated data, highlighting the potential of multimodal pre-training in advancing medical representation learning.
arXiv Detail & Related papers (2024-02-03T05:48:50Z) - Decoding visual brain representations from electroencephalography
through Knowledge Distillation and latent diffusion models [0.12289361708127873]
We present an innovative method that employs to classify and reconstruct images from the ImageNet dataset using electroencephalography (EEG) data.
We analyzed EEG recordings from 6 participants, each exposed to 50 images spanning 40 unique semantic categories.
We incorporated an image reconstruction mechanism based on pre-trained latent diffusion models, which allowed us to generate an estimate of the images which had elicited EEG activity.
arXiv Detail & Related papers (2023-09-08T09:13:50Z) - Controllable Mind Visual Diffusion Model [58.83896307930354]
Brain signal visualization has emerged as an active research area, serving as a critical interface between the human visual system and computer vision models.
We propose a novel approach, referred to as Controllable Mind Visual Model Diffusion (CMVDM)
CMVDM extracts semantic and silhouette information from fMRI data using attribute alignment and assistant networks.
We then leverage a control model to fully exploit the extracted information for image synthesis, resulting in generated images that closely resemble the visual stimuli in terms of semantics and silhouette.
arXiv Detail & Related papers (2023-05-17T11:36:40Z) - See What You See: Self-supervised Cross-modal Retrieval of Visual
Stimuli from Brain Activity [37.837710340954374]
We present a single-stage EEG-visual retrieval paradigm where data of two modalities are correlated, as opposed to their annotations.
We demonstrate the proposed approach completes an instance-level EEG-visual retrieval task which existing methods cannot.
arXiv Detail & Related papers (2022-08-07T08:11:15Z) - Exploring CLIP for Assessing the Look and Feel of Images [87.97623543523858]
We introduce Contrastive Language-Image Pre-training (CLIP) models for assessing both the quality perception (look) and abstract perception (feel) of images in a zero-shot manner.
Our results show that CLIP captures meaningful priors that generalize well to different perceptual assessments.
arXiv Detail & Related papers (2022-07-25T17:58:16Z) - SEGA: Semantic Guided Attention on Visual Prototype for Few-Shot
Learning [85.2093650907943]
We propose SEmantic Guided Attention (SEGA) to teach machines to recognize a new category.
SEGA uses semantic knowledge to guide the visual perception in a top-down manner about what visual features should be paid attention to.
We show that our semantic guided attention realizes anticipated function and outperforms state-of-the-art results.
arXiv Detail & Related papers (2021-11-08T08:03:44Z) - Emotional EEG Classification using Connectivity Features and
Convolutional Neural Networks [81.74442855155843]
We introduce a new classification system that utilizes brain connectivity with a CNN and validate its effectiveness via the emotional video classification.
The level of concentration of the brain connectivity related to the emotional property of the target video is correlated with classification performance.
arXiv Detail & Related papers (2021-01-18T13:28:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.