EEGEyeNet: a Simultaneous Electroencephalography and Eye-tracking
Dataset and Benchmark for Eye Movement Prediction
- URL: http://arxiv.org/abs/2111.05100v2
- Date: Wed, 10 Nov 2021 08:22:39 GMT
- Title: EEGEyeNet: a Simultaneous Electroencephalography and Eye-tracking
Dataset and Benchmark for Eye Movement Prediction
- Authors: Ard Kastrati, Martyna Beata P{\l}omecka, Dami\'an Pascual, Lukas Wolf,
Victor Gillioz, Roger Wattenhofer, Nicolas Langer
- Abstract summary: We present a new dataset with the goal of advancing research in the intersection of brain activities and eye movements.
EEGEyeNet consists of simultaneous Electroencephalography (EEG) and Eye-tracking (ET) recordings from 356 different subjects.
We also propose a benchmark to evaluate gaze prediction from EEG measurements.
- Score: 5.10183147987411
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We present a new dataset and benchmark with the goal of advancing research in
the intersection of brain activities and eye movements. Our dataset, EEGEyeNet,
consists of simultaneous Electroencephalography (EEG) and Eye-tracking (ET)
recordings from 356 different subjects collected from three different
experimental paradigms. Using this dataset, we also propose a benchmark to
evaluate gaze prediction from EEG measurements. The benchmark consists of three
tasks with an increasing level of difficulty: left-right, angle-amplitude and
absolute position. We run extensive experiments on this benchmark in order to
provide solid baselines, both based on classical machine learning models and on
large neural networks. We release our complete code and data and provide a
simple and easy-to-use interface to evaluate new methods.
Related papers
- Machine Learning Fairness for Depression Detection using EEG Data [14.61416119202288]
This paper presents the very first attempt to evaluate machine learning fairness for depression detection using electroencephalogram (EEG) data.
We conduct experiments using different deep learning architectures such as CNN, Long Short-Term Memory (LSTM) networks, and Gated Recurrent Unit (GRU) networks.
Our experimental results show that bias exists in existing EEG datasets and algorithms for depression detection, and different bias mitigation methods address bias at different levels across different fairness measures.
arXiv Detail & Related papers (2025-01-30T08:13:01Z) - Oriented Tiny Object Detection: A Dataset, Benchmark, and Dynamic Unbiased Learning [51.170479006249195]
We introduce a new dataset, benchmark, and a dynamic coarse-to-fine learning scheme in this study.
Our proposed dataset, AI-TOD-R, features the smallest object sizes among all oriented object detection datasets.
We present a benchmark spanning a broad range of detection paradigms, including both fully-supervised and label-efficient approaches.
arXiv Detail & Related papers (2024-12-16T09:14:32Z) - Heuristical Comparison of Vision Transformers Against Convolutional Neural Networks for Semantic Segmentation on Remote Sensing Imagery [0.0]
Vision Transformers (ViT) have brought a new wave of research in the field of computer vision.
This paper focuses on the comparison of three key factors of using (or not using) ViT for semantic segmentation of aerial images.
We show that the novel combined weighted loss function significantly boosts the CNN model's performance compared to transfer learning with ViT.
arXiv Detail & Related papers (2024-11-14T00:18:04Z) - Predicting Infant Brain Connectivity with Federated Multi-Trajectory
GNNs using Scarce Data [54.55126643084341]
Existing deep learning solutions suffer from three major limitations.
We introduce FedGmTE-Net++, a federated graph-based multi-trajectory evolution network.
Using the power of federation, we aggregate local learnings among diverse hospitals with limited datasets.
arXiv Detail & Related papers (2024-01-01T10:20:01Z) - MV-JAR: Masked Voxel Jigsaw and Reconstruction for LiDAR-Based
Self-Supervised Pre-Training [58.07391711548269]
Masked Voxel Jigsaw and Reconstruction (MV-JAR) method for LiDAR-based self-supervised pre-training.
Masked Voxel Jigsaw and Reconstruction (MV-JAR) method for LiDAR-based self-supervised pre-training.
arXiv Detail & Related papers (2023-03-23T17:59:02Z) - Learning 3D Human Pose Estimation from Dozens of Datasets using a
Geometry-Aware Autoencoder to Bridge Between Skeleton Formats [80.12253291709673]
We propose a novel affine-combining autoencoder (ACAE) method to perform dimensionality reduction on the number of landmarks.
Our approach scales to an extreme multi-dataset regime, where we use 28 3D human pose datasets to supervise one model.
arXiv Detail & Related papers (2022-12-29T22:22:49Z) - Fusing Local Similarities for Retrieval-based 3D Orientation Estimation
of Unseen Objects [70.49392581592089]
We tackle the task of estimating the 3D orientation of previously-unseen objects from monocular images.
We follow a retrieval-based strategy and prevent the network from learning object-specific features.
Our experiments on the LineMOD, LineMOD-Occluded, and T-LESS datasets show that our method yields a significantly better generalization to unseen objects than previous works.
arXiv Detail & Related papers (2022-03-16T08:53:00Z) - Facial Age Estimation using Convolutional Neural Networks [0.0]
This paper is a part of a student project in Machine Learning at the Norwegian University of Science and Technology.
A deep convolutional neural network with five convolutional layers and three fully-connected layers is presented to estimate the ages of individuals based on images.
arXiv Detail & Related papers (2021-05-14T10:09:47Z) - Deep learning-based classification of fine hand movements from low
frequency EEG [5.414308305392762]
The classification of different fine hand movements from EEG signals represents a relevant research challenge.
We trained and tested a newly proposed convolutional neural network (CNN)
CNN achieved good performance in both datasets and they were similar or superior to the baseline models.
arXiv Detail & Related papers (2020-11-13T07:16:06Z) - Benchmarking Unsupervised Object Representations for Video Sequences [111.81492107649889]
We compare the perceptual abilities of four object-centric approaches: ViMON, OP3, TBA and SCALOR.
Our results suggest that the architectures with unconstrained latent representations learn more powerful representations in terms of object detection, segmentation and tracking.
Our benchmark may provide fruitful guidance towards learning more robust object-centric video representations.
arXiv Detail & Related papers (2020-06-12T09:37:24Z) - Denoising IMU Gyroscopes with Deep Learning for Open-Loop Attitude
Estimation [0.0]
This paper proposes a learning method for denoising gyroscopes of Inertial Measurement Units (IMUs) using ground truth data.
The obtained algorithm outperforms the state-of-the-art on the (unseen) test sequences.
arXiv Detail & Related papers (2020-02-25T08:04:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.