Dual-Mode Visual System for Brain-Computer Interfaces: Integrating SSVEP and P300 Responses
- URL: http://arxiv.org/abs/2509.15439v1
- Date: Thu, 18 Sep 2025 21:25:18 GMT
- Title: Dual-Mode Visual System for Brain-Computer Interfaces: Integrating SSVEP and P300 Responses
- Authors: Ekgari Kasawala, Surej Mouli,
- Abstract summary: This study presents the development and evaluation of a novel light-emitting diode (LED)-based dual stimulation apparatus.<n>The system employs four distinct frequencies, 7 Hz, 8 Hz, 9 Hz, and 10 Hz, corresponding to forward, backward, right, and left directional controls.<n>The proposed hybrid system achieved a mean classification accuracy of 86.25%, coupled with an average ITR of 42.08 bits per minute (bpm)
- Score: 0.5801044612920815
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In brain-computer interface (BCI) systems, steady-state visual evoked potentials (SSVEP) and P300 responses have achieved widespread implementation owing to their superior information transfer rates (ITR) and minimal training requirements. These neurophysiological signals have exhibited robust efficacy and versatility in external device control, demonstrating enhanced precision and scalability. However, conventional implementations predominantly utilise liquid crystal display (LCD)-based visual stimulation paradigms, which present limitations in practical deployment scenarios. This investigation presents the development and evaluation of a novel light-emitting diode (LED)-based dual stimulation apparatus designed to enhance SSVEP classification accuracy through the integration of both SSVEP and P300 paradigms. The system employs four distinct frequencies, 7 Hz, 8 Hz, 9 Hz, and 10 Hz, corresponding to forward, backward, right, and left directional controls, respectively. Oscilloscopic verification confirmed the precision of these stimulation frequencies. Real-time feature extraction was accomplished through the concurrent analysis of maximum Fast Fourier Transform (FFT) amplitude and P300 peak detection to ascertain user intent. Directional control was determined by the frequency exhibiting maximal amplitude characteristics. The visual stimulation hardware demonstrated minimal frequency deviation, with error differentials ranging from 0.15%to 0.20%across all frequencies. The implemented signal processing algorithm successfully discriminated all four stimulus frequencies whilst correlating them with their respective P300 event markers. Classification accuracy was evaluated based on correct task intention recognition. The proposed hybrid system achieved a mean classification accuracy of 86.25%, coupled with an average ITR of 42.08 bits per minute (bpm).
Related papers
- Sparse Bayesian Modeling of EEG Channel Interactions Improves P300 Brain-Computer Interface Performance [3.9812825396617644]
EEG-based P300 brain-computer interfaces (BCIs) enable communication without physical movement by detecting stimulus-evoked neural responses.<n> Accurate and efficient decoding remains challenging due to high dimensionality, temporal dependence, and complex interactions across EEG channels.<n>We propose a sparse Bayesian time-varying regression framework that explicitly models structured EEG channel interactions.
arXiv Detail & Related papers (2026-02-19T19:03:51Z) - PRISM: Performer RS-IMLE for Single-pass Multisensory Imitation Learning [51.24484551729328]
We introduce PRISM, a single-pass policy based on a batch-global rejection-sampling variant of IMLE.<n> PRISM couples a temporal multisensory encoder with a linear-attention generator using a Performer architecture.<n>We demonstrate the efficacy of PRISM on a diverse real-world hardware suite, including loco-manipulation using a Unitree Go2 with a 7-DoF arm D1 and tabletop manipulation with a UR5 manipulator.
arXiv Detail & Related papers (2026-02-02T17:57:37Z) - A Sensing Dataset Protocol for Benchmarking and Multi-Task Wireless Sensing [45.99328292838584]
We propose the Sensing dataset Protocol (SDP), a protocol-level specification and benchmark framework for large-scale wireless sensing.<n>SDP defines how heterogeneous wireless signals are mapped into a unified perception data-block schema.<n>A unified benchmark is established for detection, recognition, and vital-sign estimation with consistent preprocessing, training, and evaluation.
arXiv Detail & Related papers (2025-12-13T05:01:30Z) - Know Me by My Pulse: Toward Practical Continuous Authentication on Wearable Devices via Wrist-Worn PPG [11.399895387051329]
We present the first real-world implementation of a continuous authentication system on a smartwatch, We-Be Band, using low-frequency (25 Hz) multi-channel PPG signals.<n>Our method employs a Bi-LSTM with attention mechanism to extract identity-specific features from short (4 s) windows of 4-channel PPG.<n>We find that sampling at 25 Hz preserves authentication accuracy, whereas performance drops sharply at 20 without compromising performance.
arXiv Detail & Related papers (2025-08-19T09:46:01Z) - Neuromorphic Wireless Split Computing with Resonate-and-Fire Neurons [69.73249913506042]
This paper investigates a wireless split computing architecture that employs resonate-and-fire (RF) neurons to process time-domain signals directly.<n>By resonating at tunable frequencies, RF neurons extract time-localized spectral features while maintaining low spiking activity.<n> Experimental results show that the proposed RF-SNN architecture achieves comparable accuracy to conventional LIF-SNNs and ANNs.
arXiv Detail & Related papers (2025-06-24T21:14:59Z) - Person Recognition at Altitude and Range: Fusion of Face, Body Shape and Gait [70.00430652562012]
FarSight is an end-to-end system for person recognition that integrates biometric cues across face, gait, and body shape modalities.<n>FarSight incorporates novel algorithms across four core modules: multi-subject detection and tracking, recognition-aware video restoration, modality-specific biometric feature encoding, and quality-guided multi-modal fusion.
arXiv Detail & Related papers (2025-05-07T17:58:25Z) - Frequency Domain Enhanced U-Net for Low-Frequency Information-Rich Image Segmentation in Surgical and Deep-Sea Exploration Robots [34.28684917337352]
We address the differences in frequency band sensitivity between CNNs and the human visual system.<n>We propose a wavelet adaptive spectrum fusion (WASF) method inspired by biological vision mechanisms to balance cross-frequency image features.<n>We develop the FE-UNet model, which employs a SAM2 backbone network and incorporates fine-tuned Hiera-Large modules to ensure segmentation accuracy.
arXiv Detail & Related papers (2025-02-06T07:24:34Z) - MLCA-AVSR: Multi-Layer Cross Attention Fusion based Audio-Visual Speech Recognition [62.89464258519723]
We propose a multi-layer cross-attention fusion based AVSR approach that promotes representation of each modality by fusing them at different levels of audio/visual encoders.
Our proposed approach surpasses the first-place system, establishing a new SOTA cpCER of 29.13% on this dataset.
arXiv Detail & Related papers (2024-01-07T08:59:32Z) - Energy-Efficient Tree-Based EEG Artifact Detection [17.085570466000906]
In epilepsy monitoring, EEG artifacts are often mistaken for seizures due to their morphological similarity in both amplitude and frequency.
In this work we present the implementation of an artifact detection algorithm based on a minimal number of EEG channels on a parallel ultra-low-power (PULP) embedded platform.
arXiv Detail & Related papers (2022-04-19T12:57:26Z) - Improving Generalization of Deep Neural Network Acoustic Models with
Length Perturbation and N-best Based Label Smoothing [49.82147684491619]
We introduce two techniques to improve generalization of deep neural network (DNN) acoustic models for automatic speech recognition (ASR)
Length perturbation is a data augmentation algorithm that randomly drops and inserts frames of an utterance to alter the length of the speech feature sequence.
N-best based label smoothing randomly injects noise to ground truth labels during training in order to avoid overfitting, where the noisy labels are generated from n-best hypotheses.
arXiv Detail & Related papers (2022-03-29T01:40:22Z) - Low Latency Real-Time Seizure Detection Using Transfer Deep Learning [0.0]
Scalp electroencephalogram (EEG) signals inherently have a low signal-to-noise ratio.
Most popular approaches to seizure detection using deep learning do not jointly model this information or require multiple passes over the signal.
In this paper, we exploit both simultaneously by converting the multichannel signal to a grayscale image and using transfer learning to achieve high performance.
arXiv Detail & Related papers (2022-02-16T00:03:00Z) - Towards Long-term Non-invasive Monitoring for Epilepsy via Wearable EEG
Devices [11.622034020961912]
We present the implementation of seizure detection algorithms based on a minimal number of EEG channels on a parallel ultra-low-power embedded platform.
We analyze global and subject-specific approaches, considering all 23-electrodes or only 4 temporal channels.
For 8s window size and subject-specific approach, we report zero false positives and 100% sensitivity.
arXiv Detail & Related papers (2021-06-15T09:37:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.