Non-contact Real time Eye Gaze Mapping System Based on Deep
Convolutional Neural Network
- URL: http://arxiv.org/abs/2009.04645v1
- Date: Thu, 10 Sep 2020 02:37:37 GMT
- Title: Non-contact Real time Eye Gaze Mapping System Based on Deep
Convolutional Neural Network
- Authors: Hoyeon Ahn
- Abstract summary: We propose a non-contact gaze mapping system applicable in real-world environments.
We introduce the GIST Gaze Mapping dataset, a Gaze mapping dataset created to learn and evaluate gaze mapping.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Human-Computer Interaction(HCI) is a field that studies interactions between
human users and computer systems. With the development of HCI, individuals or
groups of people can use various digital technologies to achieve the optimal
user experience. Human visual attention and visual intelligence are related to
cognitive science, psychology, and marketing informatics, and are used in
various applications of HCI. Gaze recognition is closely related to the HCI
field because it is meaningful in that it can enhance understanding of basic
human behavior. We can obtain reliable visual attention by the Gaze Matching
method that finds the area the user is staring at. In the previous methods, the
user wears a glasses-type device which in the form of glasses equipped with a
gaze tracking function and performs gaze tracking within a limited monitor
area. Also, the gaze estimation within a limited range is performed while the
user's posture is fixed. We overcome the physical limitations of the previous
method in this paper and propose a non-contact gaze mapping system applicable
in real-world environments. In addition, we introduce the GIST Gaze Mapping
(GGM) dataset, a Gaze mapping dataset created to learn and evaluate gaze
mapping.
Related papers
- Modeling User Preferences via Brain-Computer Interfacing [54.3727087164445]
We use Brain-Computer Interfacing technology to infer users' preferences, their attentional correlates towards visual content, and their associations with affective experience.
We link these to relevant applications, such as information retrieval, personalized steering of generative models, and crowdsourcing population estimates of affective experiences.
arXiv Detail & Related papers (2024-05-15T20:41:46Z) - Low-cost Geometry-based Eye Gaze Detection using Facial Landmarks
Generated through Deep Learning [0.0937465283958018]
We leverage novel face landmark detection neural networks to generate accurate and stable 3D landmarks of the face and iris.
Our approach demonstrates the ability to predict gaze with an angular error of less than 1.9 degrees, rivaling state-of-the-art systems.
arXiv Detail & Related papers (2023-12-31T05:45:22Z) - Pose2Gaze: Eye-body Coordination during Daily Activities for Gaze Prediction from Full-body Poses [11.545286742778977]
We first report a comprehensive analysis of eye-body coordination in various human-object and human-human interaction activities.
We then present Pose2Gaze, a eye-body coordination model that uses a convolutional neural network to extract features from head direction and full-body poses.
arXiv Detail & Related papers (2023-12-19T10:55:46Z) - CLERA: A Unified Model for Joint Cognitive Load and Eye Region Analysis
in the Wild [18.79132232751083]
Real-time analysis of the dynamics of the eye region allows us to monitor humans' visual attention allocation and estimate their mental state.
We propose CLERA, which achieves precise keypoint detection andtemporal tracking in a joint-learning framework.
We also introduce a large-scale dataset of 30k human faces with joint pupil, eye-openness, and landmark annotation.
arXiv Detail & Related papers (2023-06-26T21:20:23Z) - Leveraging Human Selective Attention for Medical Image Analysis with
Limited Training Data [72.1187887376849]
The selective attention mechanism helps the cognition system focus on task-relevant visual clues by ignoring the presence of distractors.
We propose a framework to leverage gaze for medical image analysis tasks with small training data.
Our method is demonstrated to achieve superior performance on both 3D tumor segmentation and 2D chest X-ray classification tasks.
arXiv Detail & Related papers (2021-12-02T07:55:25Z) - Understanding Character Recognition using Visual Explanations Derived
from the Human Visual System and Deep Networks [6.734853055176694]
We examine the congruence, or lack thereof, in the information-gathering strategies of deep neural networks.
The deep learning model considered similar regions in character, which humans have fixated in the case of correctly classified characters.
We propose to use the visual fixation maps obtained from the eye-tracking experiment as a supervisory input to align the model's focus on relevant character regions.
arXiv Detail & Related papers (2021-08-10T10:09:37Z) - AEGIS: A real-time multimodal augmented reality computer vision based
system to assist facial expression recognition for individuals with autism
spectrum disorder [93.0013343535411]
This paper presents the development of a multimodal augmented reality (AR) system which combines the use of computer vision and deep convolutional neural networks (CNN)
The proposed system, which we call AEGIS, is an assistive technology deployable on a variety of user devices including tablets, smartphones, video conference systems, or smartglasses.
We leverage both spatial and temporal information in order to provide an accurate expression prediction, which is then converted into its corresponding visualization and drawn on top of the original video frame.
arXiv Detail & Related papers (2020-10-22T17:20:38Z) - What Can You Learn from Your Muscles? Learning Visual Representation
from Human Interactions [50.435861435121915]
We use human interaction and attention cues to investigate whether we can learn better representations compared to visual-only representations.
Our experiments show that our "muscly-supervised" representation outperforms a visual-only state-of-the-art method MoCo.
arXiv Detail & Related papers (2020-10-16T17:46:53Z) - Towards Hardware-Agnostic Gaze-Trackers [0.5512295869673146]
We present a deep neural network architecture as an appearance-based method for constrained gaze-tracking.
Our system achieved an error of 1.8073cm on GazeCapture dataset without any calibration or device specific fine-tuning.
arXiv Detail & Related papers (2020-10-11T00:53:57Z) - Semantics-aware Adaptive Knowledge Distillation for Sensor-to-Vision
Action Recognition [131.6328804788164]
We propose a framework, named Semantics-aware Adaptive Knowledge Distillation Networks (SAKDN), to enhance action recognition in vision-sensor modality (videos)
The SAKDN uses multiple wearable-sensors as teacher modalities and uses RGB videos as student modality.
arXiv Detail & Related papers (2020-09-01T03:38:31Z) - Continuous Emotion Recognition via Deep Convolutional Autoencoder and
Support Vector Regressor [70.2226417364135]
It is crucial that the machine should be able to recognize the emotional state of the user with high accuracy.
Deep neural networks have been used with great success in recognizing emotions.
We present a new model for continuous emotion recognition based on facial expression recognition.
arXiv Detail & Related papers (2020-01-31T17:47:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.