Speech Recognition using EEG signals recorded using dry electrodes
- URL: http://arxiv.org/abs/2008.07621v1
- Date: Thu, 13 Aug 2020 09:56:45 GMT
- Title: Speech Recognition using EEG signals recorded using dry electrodes
- Authors: Gautam Krishna, Co Tran, Mason Carnahan, Morgan M Hagood, Ahmed H
Tewfik
- Abstract summary: We demonstrate speech recognition using electroencephalography (EEG) signals obtained using dry electrodes.
We demonstrate a test accuracy of 79.07 percent on a subset vocabulary consisting of two English vowels.
- Score: 12.417540155936717
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this paper, we demonstrate speech recognition using electroencephalography
(EEG) signals obtained using dry electrodes on a limited English vocabulary
consisting of three vowels and one word using a deep learning model. We
demonstrate a test accuracy of 79.07 percent on a subset vocabulary consisting
of two English vowels. Our results demonstrate the feasibility of using EEG
signals recorded using dry electrodes for performing the task of speech
recognition.
Related papers
- NeuroSpex: Neuro-Guided Speaker Extraction with Cross-Modal Attention [47.8479647938849]
We present a neuro-guided speaker extraction model, i.e. NeuroSpex, using the EEG response of the listener as the sole auxiliary reference cue.
We propose a novel EEG signal encoder that captures the attention information. Additionally, we propose a cross-attention (CA) mechanism to enhance the speech feature representations.
arXiv Detail & Related papers (2024-09-04T07:33:01Z) - Enhancing EEG-to-Text Decoding through Transferable Representations from Pre-trained Contrastive EEG-Text Masked Autoencoder [69.7813498468116]
We propose Contrastive EEG-Text Masked Autoencoder (CET-MAE), a novel model that orchestrates compound self-supervised learning across and within EEG and text.
We also develop a framework called E2T-PTR (EEG-to-Text decoding using Pretrained Transferable Representations) to decode text from EEG sequences.
arXiv Detail & Related papers (2024-02-27T11:45:21Z) - DGSD: Dynamical Graph Self-Distillation for EEG-Based Auditory Spatial
Attention Detection [49.196182908826565]
Auditory Attention Detection (AAD) aims to detect target speaker from brain signals in a multi-speaker environment.
Current approaches primarily rely on traditional convolutional neural network designed for processing Euclidean data like images.
This paper proposes a dynamical graph self-distillation (DGSD) approach for AAD, which does not require speech stimuli as input.
arXiv Detail & Related papers (2023-09-07T13:43:46Z) - NUVA: A Naming Utterance Verifier for Aphasia Treatment [49.114436579008476]
Assessment of speech performance using picture naming tasks is a key method for both diagnosis and monitoring of responses to treatment interventions by people with aphasia (PWA)
Here we present NUVA, an utterance verification system incorporating a deep learning element that classifies 'correct' versus'incorrect' naming attempts from aphasic stroke patients.
When tested on eight native British-English speaking PWA the system's performance accuracy ranged between 83.6% to 93.6%, with a 10-fold cross-validation mean of 89.5%.
arXiv Detail & Related papers (2021-02-10T13:00:29Z) - Pre-Training Transformers as Energy-Based Cloze Models [95.04748595976811]
We introduce Electric, an energy-based cloze model for representation learning over text.
Electric does not use masking or output a full distribution over tokens that could occur in a context.
We train Electric using an algorithm based on noise-contrastive estimation and elucidate how this learning objective is closely related to the recently proposed ELECTRA pre-training method.
arXiv Detail & Related papers (2020-12-15T19:17:33Z) - Digital Voicing of Silent Speech [48.15708685020142]
We consider the task of digitally voicing silent speech, where silently mouthed words are converted to audible speech based on electromyography (EMG) sensor measurements.
We introduce a method of training on silent EMG by transferring audio targets from vocalized to silent signals.
Our method greatly improves intelligibility of audio generated from silent EMG compared to a baseline that only trains with vocalized data.
arXiv Detail & Related papers (2020-10-06T18:23:35Z) - Constrained Variational Autoencoder for improving EEG based Speech
Recognition Systems [3.5786621294068377]
We introduce a recurrent neural network (RNN) based variational autoencoder (VAE) model with a new constrained loss function.
We demonstrate that both continuous and isolated speech recognition systems trained and tested using EEG features generated from raw EEG features.
arXiv Detail & Related papers (2020-06-01T06:03:50Z) - Improving EEG based continuous speech recognition using GAN [3.5786621294068377]
We demonstrate that it is possible to generate more meaningful electroencephalography (EEG) features from raw EEG features using generative adversarial networks (GAN)
Our proposed approach can be implemented without using any additional sensor information, whereas in [1] authors used additional features like acoustic or articulatory information to improve the performance of EEG based continuous speech recognition systems.
arXiv Detail & Related papers (2020-05-29T06:11:33Z) - Understanding effect of speech perception in EEG based speech
recognition systems [3.5786621294068377]
The electroencephalography (EEG) signals recorded in parallel with speech are used to perform isolated and continuous speech recognition.
We investigate whether it is possible to separate out this speech perception component from EEG signals in order to design more robust EEG based speech recognition systems.
arXiv Detail & Related papers (2020-05-29T05:56:09Z) - Speech Synthesis using EEG [4.312746668772343]
We make use of a recurrent neural network (RNN) regression model to predict acoustic features directly from EEG features.
We provide EEG based speech synthesis results for four subjects in this paper.
arXiv Detail & Related papers (2020-02-22T03:53:45Z) - Continuous Silent Speech Recognition using EEG [3.5786621294068377]
We translate EEG signals recorded in parallel while subjects were reading English sentences in their mind without producing any voice to text.
Our results demonstrate the feasibility of using EEG signals for performing continuous silent speech recognition.
arXiv Detail & Related papers (2020-02-06T18:28:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.