CLRGaze: Contrastive Learning of Representations for Eye Movement
Signals
- URL: http://arxiv.org/abs/2010.13046v2
- Date: Sun, 30 May 2021 14:14:23 GMT
- Title: CLRGaze: Contrastive Learning of Representations for Eye Movement
Signals
- Authors: Louise Gillian C. Bautista and Prospero C. Naval Jr
- Abstract summary: We learn feature vectors of eye movements in a self-supervised manner.
We adopt a contrastive learning approach and propose a set of data transformations that encourage a deep neural network to discern salient and granular gaze patterns.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Eye movements are intricate and dynamic biosignals that contain a wealth of
cognitive information about the subject. However, these are ambiguous signals
and therefore require meticulous feature engineering to be used by machine
learning algorithms. We instead propose to learn feature vectors of eye
movements in a self-supervised manner. We adopt a contrastive learning approach
and propose a set of data transformations that encourage a deep neural network
to discern salient and granular gaze patterns. This paper presents a novel
experiment utilizing six eye-tracking data sets despite different data
specifications and experimental conditions. We assess the learned features on
biometric tasks with only a linear classifier, achieving 84.6% accuracy on a
mixed dataset, and up to 97.3% accuracy on a single dataset. Our work advances
the state of machine learning for eye movements and provides insights into a
general representation learning method not only for eye movements but also for
similar biosignals.
Related papers
- Deep Domain Adaptation: A Sim2Real Neural Approach for Improving Eye-Tracking Systems [80.62854148838359]
Eye image segmentation is a critical step in eye tracking that has great influence over the final gaze estimate.
We use dimensionality-reduction techniques to measure the overlap between the target eye images and synthetic training data.
Our methods result in robust, improved performance when tackling the discrepancy between simulation and real-world data samples.
arXiv Detail & Related papers (2024-03-23T22:32:06Z) - What Makes Pre-Trained Visual Representations Successful for Robust
Manipulation? [57.92924256181857]
We find that visual representations designed for manipulation and control tasks do not necessarily generalize under subtle changes in lighting and scene texture.
We find that emergent segmentation ability is a strong predictor of out-of-distribution generalization among ViT models.
arXiv Detail & Related papers (2023-11-03T18:09:08Z) - URLOST: Unsupervised Representation Learning without Stationarity or
Topology [26.17135629579595]
We introduce a novel framework that learns from high-dimensional data lacking stationarity and topology.
Our model combines a learnable self-organizing layer, density adjusted spectral clustering, and masked autoencoders.
We evaluate its effectiveness on simulated biological vision data, neural recordings from the primary visual cortex, and gene expression datasets.
arXiv Detail & Related papers (2023-10-06T18:00:02Z) - Factors of Influence for Transfer Learning across Diverse Appearance
Domains and Task Types [50.1843146606122]
A simple form of transfer learning is common in current state-of-the-art computer vision models.
Previous systematic studies of transfer learning have been limited and the circumstances in which it is expected to work are not fully understood.
In this paper we carry out an extensive experimental exploration of transfer learning across vastly different image domains.
arXiv Detail & Related papers (2021-03-24T16:24:20Z) - Dynamic Graph Modeling of Simultaneous EEG and Eye-tracking Data for
Reading Task Identification [79.41619843969347]
We present a new approach, that we call AdaGTCN, for identifying human reader intent from Electroencephalogram(EEG) and Eye movement(EM) data.
Our method, Adaptive Graph Temporal Convolution Network (AdaGTCN), uses an Adaptive Graph Learning Layer and Deep Neighborhood Graph Convolution Layer.
We compare our approach with several baselines to report an improvement of 6.29% on the ZuCo 2.0 dataset, along with extensive ablation experiments.
arXiv Detail & Related papers (2021-02-21T18:19:49Z) - Relational Graph Learning on Visual and Kinematics Embeddings for
Accurate Gesture Recognition in Robotic Surgery [84.73764603474413]
We propose a novel online approach of multi-modal graph network (i.e., MRG-Net) to dynamically integrate visual and kinematics information.
The effectiveness of our method is demonstrated with state-of-the-art results on the public JIGSAWS dataset.
arXiv Detail & Related papers (2020-11-03T11:00:10Z) - Malicious Network Traffic Detection via Deep Learning: An Information
Theoretic View [0.0]
We study how homeomorphism affects learned representation of a malware traffic dataset.
Our results suggest that although the details of learned representations and the specific coordinate system defined over the manifold of all parameters differ slightly, the functional approximations are the same.
arXiv Detail & Related papers (2020-09-16T15:37:44Z) - GazeMAE: General Representations of Eye Movements using a Micro-Macro
Autoencoder [0.0]
We propose an abstract representation of eye movements that preserve the important nuances in gaze behavior while being stimuli-agnostic.
We consider eye movements as raw position and velocity signals and train separate deep temporal convolutional autoencoders.
The autoencoders learn micro-scale and macro-scale representations that correspond to the fast and slow features of eye movements.
arXiv Detail & Related papers (2020-09-05T02:13:42Z) - Semantics-aware Adaptive Knowledge Distillation for Sensor-to-Vision
Action Recognition [131.6328804788164]
We propose a framework, named Semantics-aware Adaptive Knowledge Distillation Networks (SAKDN), to enhance action recognition in vision-sensor modality (videos)
The SAKDN uses multiple wearable-sensors as teacher modalities and uses RGB videos as student modality.
arXiv Detail & Related papers (2020-09-01T03:38:31Z) - TactileSGNet: A Spiking Graph Neural Network for Event-based Tactile
Object Recognition [17.37142241982902]
New advances in flexible, event-driven, electronic skins may soon endow robots with touch perception capabilities similar to humans.
These unique features may render current deep learning approaches such as convolutional feature extractors unsuitable for tactile learning.
We propose a novel spiking graph neural network for event-based tactile object recognition.
arXiv Detail & Related papers (2020-08-01T03:35:15Z) - MLGaze: Machine Learning-Based Analysis of Gaze Error Patterns in
Consumer Eye Tracking Systems [0.0]
In this study, gaze error patterns produced by a commercial eye tracking device were studied with the help of machine learning algorithms.
It was seen that while the impact of the different error sources on gaze data characteristics were nearly impossible to distinguish by visual inspection or from data statistics, machine learning models were successful in identifying the impact of the different error sources and predicting the variability in gaze error levels due to these conditions.
arXiv Detail & Related papers (2020-05-07T23:07:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.