EEGEyeNet: a Simultaneous Electroencephalography and Eye-tracking
Dataset and Benchmark for Eye Movement Prediction
- URL: http://arxiv.org/abs/2111.05100v2
- Date: Wed, 10 Nov 2021 08:22:39 GMT
- Title: EEGEyeNet: a Simultaneous Electroencephalography and Eye-tracking
Dataset and Benchmark for Eye Movement Prediction
- Authors: Ard Kastrati, Martyna Beata P{\l}omecka, Dami\'an Pascual, Lukas Wolf,
Victor Gillioz, Roger Wattenhofer, Nicolas Langer
- Abstract summary: We present a new dataset with the goal of advancing research in the intersection of brain activities and eye movements.
EEGEyeNet consists of simultaneous Electroencephalography (EEG) and Eye-tracking (ET) recordings from 356 different subjects.
We also propose a benchmark to evaluate gaze prediction from EEG measurements.
- Score: 5.10183147987411
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We present a new dataset and benchmark with the goal of advancing research in
the intersection of brain activities and eye movements. Our dataset, EEGEyeNet,
consists of simultaneous Electroencephalography (EEG) and Eye-tracking (ET)
recordings from 356 different subjects collected from three different
experimental paradigms. Using this dataset, we also propose a benchmark to
evaluate gaze prediction from EEG measurements. The benchmark consists of three
tasks with an increasing level of difficulty: left-right, angle-amplitude and
absolute position. We run extensive experiments on this benchmark in order to
provide solid baselines, both based on classical machine learning models and on
large neural networks. We release our complete code and data and provide a
simple and easy-to-use interface to evaluate new methods.
Related papers
- HOIGaze: Gaze Estimation During Hand-Object Interactions in Extended Reality Exploiting Eye-Hand-Head Coordination [10.982807572404166]
HOIGaze is a learning-based approach for gaze estimation during hand-object interactions (HOIs) in extended reality (XR)
The eye, hand, and head movements are closely coordinated during HOIs and this coordination can be exploited to identify samples that are most useful for gaze training.
We evaluate HOIGaze on the HOT3D and Aria digital twin (ADT) datasets and show that it significantly outperforms state-of-the-art methods.
arXiv Detail & Related papers (2025-04-28T14:31:43Z) - IAAO: Interactive Affordance Learning for Articulated Objects in 3D Environments [56.85804719947]
We present IAAO, a framework that builds an explicit 3D model for intelligent agents to gain understanding of articulated objects in their environment through interaction.
We first build hierarchical features and label fields for each object state using 3D Gaussian Splatting (3DGS) by distilling mask features and view-consistent labels from multi-view images.
We then perform object- and part-level queries on the 3D Gaussian primitives to identify static and articulated elements, estimating global transformations and local articulation parameters along with affordances.
arXiv Detail & Related papers (2025-04-09T12:36:48Z) - EEG-EyeTrack: A Benchmark for Time Series and Functional Data Analysis with Open Challenges and Baselines [0.0]
The contribution is twofold: first, open challenges and evaluation metrics tailored to FDA applications are proposed.
functional neural networks are used to establish baseline results for the primary regression task.
Baseline results are reported for the new dataset, based on consumer-grade hardware, and the EEGEyeNet dataset, based on research-grade hardware.
arXiv Detail & Related papers (2025-04-02T08:33:38Z) - Consumer-grade EEG-based Eye Tracking [0.0]
Electroencephalography-based eye tracking (EEG-ET) leverages eye movement artifacts in EEG signals as an alternative to camera-based tracking.
EEG-ET offers advantages such as robustness in low-light conditions and better integration with brain-computer interfaces.
We present a dataset comprising simultaneous EEG and eye-tracking recordings from 113 participants across 116 sessions.
arXiv Detail & Related papers (2025-03-18T14:53:20Z) - Heuristical Comparison of Vision Transformers Against Convolutional Neural Networks for Semantic Segmentation on Remote Sensing Imagery [0.0]
Vision Transformers (ViT) have brought a new wave of research in the field of computer vision.
This paper focuses on the comparison of three key factors of using (or not using) ViT for semantic segmentation of aerial images.
We show that the novel combined weighted loss function significantly boosts the CNN model's performance compared to transfer learning with ViT.
arXiv Detail & Related papers (2024-11-14T00:18:04Z) - Predicting Infant Brain Connectivity with Federated Multi-Trajectory
GNNs using Scarce Data [54.55126643084341]
Existing deep learning solutions suffer from three major limitations.
We introduce FedGmTE-Net++, a federated graph-based multi-trajectory evolution network.
Using the power of federation, we aggregate local learnings among diverse hospitals with limited datasets.
arXiv Detail & Related papers (2024-01-01T10:20:01Z) - MV-JAR: Masked Voxel Jigsaw and Reconstruction for LiDAR-Based
Self-Supervised Pre-Training [58.07391711548269]
Masked Voxel Jigsaw and Reconstruction (MV-JAR) method for LiDAR-based self-supervised pre-training.
Masked Voxel Jigsaw and Reconstruction (MV-JAR) method for LiDAR-based self-supervised pre-training.
arXiv Detail & Related papers (2023-03-23T17:59:02Z) - Learning 3D Human Pose Estimation from Dozens of Datasets using a
Geometry-Aware Autoencoder to Bridge Between Skeleton Formats [80.12253291709673]
We propose a novel affine-combining autoencoder (ACAE) method to perform dimensionality reduction on the number of landmarks.
Our approach scales to an extreme multi-dataset regime, where we use 28 3D human pose datasets to supervise one model.
arXiv Detail & Related papers (2022-12-29T22:22:49Z) - Stacking Ensemble Learning in Deep Domain Adaptation for Ophthalmic
Image Classification [61.656149405657246]
Domain adaptation is effective in image classification tasks where obtaining sufficient label data is challenging.
We propose a novel method, named SELDA, for stacking ensemble learning via extending three domain adaptation methods.
The experimental results using Age-Related Eye Disease Study (AREDS) benchmark ophthalmic dataset demonstrate the effectiveness of the proposed model.
arXiv Detail & Related papers (2022-09-27T14:19:00Z) - EEG-BBNet: a Hybrid Framework for Brain Biometric using Graph
Connectivity [1.1498015270151059]
We present EEG-BBNet, a hybrid network which integrates convolutional neural networks (CNN) with graph convolutional neural networks (GCNN)
Our models outperform all baselines in the event-related potential (ERP) task with an average correct recognition rates up to 99.26% using intra-session data.
arXiv Detail & Related papers (2022-08-17T10:18:22Z) - Fusing Local Similarities for Retrieval-based 3D Orientation Estimation
of Unseen Objects [70.49392581592089]
We tackle the task of estimating the 3D orientation of previously-unseen objects from monocular images.
We follow a retrieval-based strategy and prevent the network from learning object-specific features.
Our experiments on the LineMOD, LineMOD-Occluded, and T-LESS datasets show that our method yields a significantly better generalization to unseen objects than previous works.
arXiv Detail & Related papers (2022-03-16T08:53:00Z) - Facial Age Estimation using Convolutional Neural Networks [0.0]
This paper is a part of a student project in Machine Learning at the Norwegian University of Science and Technology.
A deep convolutional neural network with five convolutional layers and three fully-connected layers is presented to estimate the ages of individuals based on images.
arXiv Detail & Related papers (2021-05-14T10:09:47Z) - Deep learning-based classification of fine hand movements from low
frequency EEG [5.414308305392762]
The classification of different fine hand movements from EEG signals represents a relevant research challenge.
We trained and tested a newly proposed convolutional neural network (CNN)
CNN achieved good performance in both datasets and they were similar or superior to the baseline models.
arXiv Detail & Related papers (2020-11-13T07:16:06Z) - Benchmarking Unsupervised Object Representations for Video Sequences [111.81492107649889]
We compare the perceptual abilities of four object-centric approaches: ViMON, OP3, TBA and SCALOR.
Our results suggest that the architectures with unconstrained latent representations learn more powerful representations in terms of object detection, segmentation and tracking.
Our benchmark may provide fruitful guidance towards learning more robust object-centric video representations.
arXiv Detail & Related papers (2020-06-12T09:37:24Z) - Denoising IMU Gyroscopes with Deep Learning for Open-Loop Attitude
Estimation [0.0]
This paper proposes a learning method for denoising gyroscopes of Inertial Measurement Units (IMUs) using ground truth data.
The obtained algorithm outperforms the state-of-the-art on the (unseen) test sequences.
arXiv Detail & Related papers (2020-02-25T08:04:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.