Speech Artifact Removal from EEG Recordings of Spoken Word Production
with Tensor Decomposition
- URL: http://arxiv.org/abs/2206.00635v1
- Date: Wed, 1 Jun 2022 17:10:23 GMT
- Title: Speech Artifact Removal from EEG Recordings of Spoken Word Production
with Tensor Decomposition
- Authors: Holy Lovenia, Hiroki Tanaka, Sakriani Sakti, Ayu Purwarianti, and
Satoshi Nakamura
- Abstract summary: Speech artifacts contaminate electroencephalogram (EEG) signals and prevent the inspection of the underlying cognitive processes.
To fuel further EEG research with speech production, a method using three-mode tensor decomposition is proposed.
In a picture-naming task, we collected raw data with speech artifacts by placing two electrodes near the mouth to record lip EMG.
- Score: 20.397149635457346
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Research about brain activities involving spoken word production is
considerably underdeveloped because of the undiscovered characteristics of
speech artifacts, which contaminate electroencephalogram (EEG) signals and
prevent the inspection of the underlying cognitive processes. To fuel further
EEG research with speech production, a method using three-mode tensor
decomposition (time x space x frequency) is proposed to perform speech artifact
removal. Tensor decomposition enables simultaneous inspection of multiple
modes, which suits the multi-way nature of EEG data. In a picture-naming task,
we collected raw data with speech artifacts by placing two electrodes near the
mouth to record lip EMG. Based on our evaluation, which calculated the
correlation values between grand-averaged speech artifacts and the lip EMG,
tensor decomposition outperformed the former methods that were based on
independent component analysis (ICA) and blind source separation (BSS), both in
detecting speech artifact (0.985) and producing clean data (0.101). Our
proposed method correctly preserved the components unrelated to speech, which
was validated by computing the correlation value between the grand-averaged raw
data without EOG and cleaned data before the speech onset (0.92-0.94).
Related papers
- Detecting COPD Through Speech Analysis: A Dataset of Danish Speech and Machine Learning Approach [4.132109134011237]
Chronic Obstructive Pulmonary Disease (COPD) is a serious and debilitating disease affecting millions around the world.<n>Our findings support the potential of speech-based analysis as a non-invasive, remote, and scalable screening tool as part of future COPD healthcare solutions.
arXiv Detail & Related papers (2025-08-04T12:44:07Z) - Direct Dual-Energy CT Material Decomposition using Model-based Denoising Diffusion Model [105.95160543743984]
We propose a deep learning procedure called Dual-Energy Decomposition Model-based Diffusion (DEcomp-MoD) for quantitative material decomposition.<n>We show that DEcomp-MoD outperform state-of-the-art unsupervised score-based model and supervised deep learning networks.
arXiv Detail & Related papers (2025-07-24T01:00:06Z) - A Silent Speech Decoding System from EEG and EMG with Heterogenous Electrode Configurations [0.20075899678041528]
We introduce neural networks that can handle EEG/EMG with heterogeneous electrode placements.<n>We show strong performance in silent speech decoding via multi-task training on large-scale EEG/EMG datasets.
arXiv Detail & Related papers (2025-06-16T07:57:35Z) - Study of the Performance of CEEMDAN in Underdetermined Speech Separation [0.0]
The CEEMDAN algorithm is one of the modern methods used in the analysis of non-stationary signals.
This research presents a study of the effectiveness of this method in audio source separation to know the limits of its work.
arXiv Detail & Related papers (2024-11-18T06:13:51Z) - Denoising VAE as an Explainable Feature Reduction and Diagnostic Pipeline for Autism Based on Resting state fMRI [11.871709357017416]
We propose a feature reduction pipeline using resting-state fMRI data.
We used Craddock atlas and Power atlas to extract functional connectivity data from rs-fMRI.
By using a denoising variational autoencoder, our proposed pipeline further compresses the connectivity features into 5 latent Gaussian distributions.
arXiv Detail & Related papers (2024-09-30T09:38:47Z) - Enhancing Electrocardiogram Signal Analysis Using NLP-Inspired Techniques: A Novel Approach with Embedding and Self-Attention [2.7651063843287718]
We propose a novel ECG analysis technique, based on embedding and self attention, to capture the spatial as well as the temporal dependencies of the ECG data.
An accuracy of 91% was achieved with a good F1-score for all the disease classes.
arXiv Detail & Related papers (2024-07-15T12:20:15Z) - HyPoradise: An Open Baseline for Generative Speech Recognition with
Large Language Models [81.56455625624041]
We introduce the first open-source benchmark to utilize external large language models (LLMs) for ASR error correction.
The proposed benchmark contains a novel dataset, HyPoradise (HP), encompassing more than 334,000 pairs of N-best hypotheses.
LLMs with reasonable prompt and its generative capability can even correct those tokens that are missing in N-best list.
arXiv Detail & Related papers (2023-09-27T14:44:10Z) - EOG Artifact Removal from Single and Multi-channel EEG Recordings
through the combination of Long Short-Term Memory Networks and Independent
Component Analysis [0.0]
We present a novel methodology that combines a long short-term memory (LSTM)-based neural network with ICA to address the challenge of EOG artifact removal from EEG signals.
Our approach aims to accomplish two primary objectives: 1) estimate the horizontal and vertical EOG signals from the contaminated EEG data, and 2) employ ICA to eliminate the estimated EOG signals from the EEG.
arXiv Detail & Related papers (2023-08-25T13:32:28Z) - Data Augmentation for Seizure Prediction with Generative Diffusion Model [34.12334834099495]
We propose a novel diffusion-based DA method called DiffEEG.
It can fully explore data distribution and generate samples with high diversity.
With the contribution of DiffEEG, the Multi-scale CNN achieves state-of-the-art performance.
arXiv Detail & Related papers (2023-06-14T05:44:53Z) - Removal of Ocular Artifacts in EEG Using Deep Learning [0.0]
The removal of ocular artifacts is the most challenging among these artifacts.
In this study, a novel ocular artifact removal method is presented by developing bidirectional long-short term memory (BiLSTM)-based deep learning (DL) models.
Our results demonstrated the WSST-Net model significantly improves artifact removal performance compared to traditional TF and raw signal methods.
arXiv Detail & Related papers (2022-09-24T11:19:52Z) - Unsupervised Anomaly Detection in 3D Brain MRI using Deep Learning with
impured training data [53.122045119395594]
We study how unhealthy samples within the training data affect anomaly detection performance for brain MRI-scans.
We evaluate a method to identify falsely labeled samples directly during training based on the reconstruction error of the AE.
arXiv Detail & Related papers (2022-04-12T13:05:18Z) - Investigation of Data Augmentation Techniques for Disordered Speech
Recognition [69.50670302435174]
This paper investigates a set of data augmentation techniques for disordered speech recognition.
Both normal and disordered speech were exploited in the augmentation process.
The final speaker adapted system constructed using the UASpeech corpus and the best augmentation approach based on speed perturbation produced up to 2.92% absolute word error rate (WER)
arXiv Detail & Related papers (2022-01-14T17:09:22Z) - Discretization and Re-synthesis: an alternative method to solve the
Cocktail Party Problem [65.25725367771075]
This study demonstrates, for the first time, that the synthesis-based approach can also perform well on this problem.
Specifically, we propose a novel speech separation/enhancement model based on the recognition of discrete symbols.
By utilizing the synthesis model with the input of discrete symbols, after the prediction of discrete symbol sequence, each target speech could be re-synthesized.
arXiv Detail & Related papers (2021-12-17T08:35:40Z) - Adaptive Multi-View ICA: Estimation of noise levels for optimal
inference [65.94843987207445]
Adaptive multiView ICA (AVICA) is a noisy ICA model where each view is a linear mixture of shared independent sources with additive noise on the sources.
On synthetic data, AVICA yields better sources estimates than other group ICA methods thanks to its explicit MMSE estimator.
On real magnetoencephalograpy (MEG) data, we provide evidence that the decomposition is less sensitive to sampling noise and that the noise variance estimates are biologically plausible.
arXiv Detail & Related papers (2021-02-22T13:10:12Z) - Multi-Modal Detection of Alzheimer's Disease from Speech and Text [3.702631194466718]
We propose a deep learning method that utilizes speech and the corresponding transcript simultaneously to detect Alzheimer's disease (AD)
The proposed method achieves 85.3% 10-fold cross-validation accuracy when trained and evaluated on the Dementiabank Pitt corpus.
arXiv Detail & Related papers (2020-11-30T21:18:17Z) - Continuous Speech Separation with Conformer [60.938212082732775]
We use transformer and conformer in lieu of recurrent neural networks in the separation system.
We believe capturing global information with the self-attention based method is crucial for the speech separation.
arXiv Detail & Related papers (2020-08-13T09:36:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.