Towards Hardware-Agnostic Gaze-Trackers
- URL: http://arxiv.org/abs/2010.05123v1
- Date: Sun, 11 Oct 2020 00:53:57 GMT
- Title: Towards Hardware-Agnostic Gaze-Trackers
- Authors: Jatin Sharma and Jon Campbell and Pete Ansell and Jay Beavers and
Christopher O'Dowd
- Abstract summary: We present a deep neural network architecture as an appearance-based method for constrained gaze-tracking.
Our system achieved an error of 1.8073cm on GazeCapture dataset without any calibration or device specific fine-tuning.
- Score: 0.5512295869673146
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Gaze-tracking is a novel way of interacting with computers which allows new
scenarios, such as enabling people with motor-neuron disabilities to control
their computers or doctors to interact with patient information without
touching screen or keyboard. Further, there are emerging applications of
gaze-tracking in interactive gaming, user experience research, human attention
analysis and behavioral studies. Accurate estimation of the gaze may involve
accounting for head-pose, head-position, eye rotation, distance from the object
as well as operating conditions such as illumination, occlusion, background
noise and various biological aspects of the user. Commercially available
gaze-trackers utilize specialized sensor assemblies that usually consist of an
infrared light source and camera. There are several challenges in the universal
proliferation of gaze-tracking as accessibility technologies, specifically its
affordability, reliability, and ease-of-use. In this paper, we try to address
these challenges through the development of a hardware-agnostic gaze-tracker.
We present a deep neural network architecture as an appearance-based method for
constrained gaze-tracking that utilizes facial imagery captured on an ordinary
RGB camera ubiquitous in all modern computing devices. Our system achieved an
error of 1.8073cm on GazeCapture dataset without any calibration or device
specific fine-tuning. This research shows promise that one day soon any
computer, tablet, or phone will be controllable using just your eyes due to the
prediction capabilities of deep neutral networks.
Related papers
- Tri-Cam: Practical Eye Gaze Tracking via Camera Network [1.642094639107215]
We introduce Tri-Cam, a practical deep learning-based gaze tracking system using three affordable RGB webcams.
It features a split network structure for efficient training, as well as designated network designs to handle the separated gaze tracking tasks.
We evaluate Tri-Cam against Tobii, the state-of-the-art commercial eye tracker, achieving comparable accuracy, while supporting a wider free movement area.
arXiv Detail & Related papers (2024-09-29T04:43:10Z) - Floor extraction and door detection for visually impaired guidance [78.94595951597344]
Finding obstacle-free paths in unknown environments is a big navigation issue for visually impaired people and autonomous robots.
New devices based on computer vision systems can help impaired people to overcome the difficulties of navigating in unknown environments in safe conditions.
In this work it is proposed a combination of sensors and algorithms that can lead to the building of a navigation system for visually impaired people.
arXiv Detail & Related papers (2024-01-30T14:38:43Z) - Low-cost Geometry-based Eye Gaze Detection using Facial Landmarks
Generated through Deep Learning [0.0937465283958018]
We leverage novel face landmark detection neural networks to generate accurate and stable 3D landmarks of the face and iris.
Our approach demonstrates the ability to predict gaze with an angular error of less than 1.9 degrees, rivaling state-of-the-art systems.
arXiv Detail & Related papers (2023-12-31T05:45:22Z) - Open Gaze: Open Source eye tracker for smartphone devices using Deep Learning [0.0]
We present an open-source implementation of a smartphone-based gaze tracker that emulates the methodology proposed by a GooglePaper.
Through the integration of machine learning techniques, we unveil an accurate eye tracking solution that is native to smartphones.
Our findings exhibit the inherent potential to amplify eye movement research by significant proportions.
arXiv Detail & Related papers (2023-08-25T17:10:22Z) - Multimodal Adaptive Fusion of Face and Gait Features using Keyless
attention based Deep Neural Networks for Human Identification [67.64124512185087]
Soft biometrics such as gait are widely used with face in surveillance tasks like person recognition and re-identification.
We propose a novel adaptive multi-biometric fusion strategy for the dynamic incorporation of gait and face biometric cues by leveraging keyless attention deep neural networks.
arXiv Detail & Related papers (2023-03-24T05:28:35Z) - DensePose From WiFi [86.61881052177228]
We develop a deep neural network that maps the phase and amplitude of WiFi signals to UV coordinates within 24 human regions.
Our model can estimate the dense pose of multiple subjects, with comparable performance to image-based approaches.
arXiv Detail & Related papers (2022-12-31T16:48:43Z) - Force-Aware Interface via Electromyography for Natural VR/AR Interaction [69.1332992637271]
We design a learning-based neural interface for natural and intuitive force inputs in VR/AR.
We show that our interface can decode finger-wise forces in real-time with 3.3% mean error, and generalize to new users with little calibration.
We envision our findings to push forward research towards more realistic physicality in future VR/AR.
arXiv Detail & Related papers (2022-10-03T20:51:25Z) - Unmasking Communication Partners: A Low-Cost AI Solution for Digitally
Removing Head-Mounted Displays in VR-Based Telepresence [62.997667081978825]
Face-to-face conversation in Virtual Reality (VR) is a challenge when participants wear head-mounted displays (HMD)
Past research has shown that high-fidelity face reconstruction with personal avatars in VR is possible under laboratory conditions with high-cost hardware.
We propose one of the first low-cost systems for this task which uses only open source, free software and affordable hardware.
arXiv Detail & Related papers (2020-11-06T23:17:12Z) - AEGIS: A real-time multimodal augmented reality computer vision based
system to assist facial expression recognition for individuals with autism
spectrum disorder [93.0013343535411]
This paper presents the development of a multimodal augmented reality (AR) system which combines the use of computer vision and deep convolutional neural networks (CNN)
The proposed system, which we call AEGIS, is an assistive technology deployable on a variety of user devices including tablets, smartphones, video conference systems, or smartglasses.
We leverage both spatial and temporal information in order to provide an accurate expression prediction, which is then converted into its corresponding visualization and drawn on top of the original video frame.
arXiv Detail & Related papers (2020-10-22T17:20:38Z) - Non-contact Real time Eye Gaze Mapping System Based on Deep
Convolutional Neural Network [0.0]
We propose a non-contact gaze mapping system applicable in real-world environments.
We introduce the GIST Gaze Mapping dataset, a Gaze mapping dataset created to learn and evaluate gaze mapping.
arXiv Detail & Related papers (2020-09-10T02:37:37Z) - Automated analysis of eye-tracker-based human-human interaction studies [2.433293618209319]
We investigate which state-of-the-art computer vision algorithms may be used to automate the post-analysis of mobile eye-tracking data.
For the case study in this paper, we focus on mobile eye-tracker recordings made during human-human face-to-face interactions.
We show that the use of this single-pipeline framework provides robust results, which are both more accurate and faster than previous work in the field.
arXiv Detail & Related papers (2020-07-09T10:00:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.