AutoSIGHT: Automatic Eye Tracking-based System for Immediate Grading of Human experTise
- URL: http://arxiv.org/abs/2508.01015v1
- Date: Fri, 01 Aug 2025 18:28:13 GMT
- Title: AutoSIGHT: Automatic Eye Tracking-based System for Immediate Grading of Human experTise
- Authors: Byron Dowling, Jozef Probcin, Adam Czajka,
- Abstract summary: This paper proposes AutoSIGHT, Automatic System for Immediate Grading of Human experTise, that classifies expert and non-expert performers.<n>Results on the task of iris Presentation Attack Detection (PAD) used for this study show that with a small evaluation window of just 5 seconds, AutoSIGHT achieves an average average Area Under the ROC curve performance of 0.751.<n>This work opens new areas of research on how to incorporate the automatic weighing of human and machine expertise into human-AI pairing setups.
- Score: 4.215251065887862
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Can we teach machines to assess the expertise of humans solving visual tasks automatically based on eye tracking features? This paper proposes AutoSIGHT, Automatic System for Immediate Grading of Human experTise, that classifies expert and non-expert performers, and builds upon an ensemble of features extracted from eye tracking data while the performers were solving a visual task. Results on the task of iris Presentation Attack Detection (PAD) used for this study show that with a small evaluation window of just 5 seconds, AutoSIGHT achieves an average average Area Under the ROC curve performance of 0.751 in subject-disjoint train-test regime, indicating that such detection is viable. Furthermore, when a larger evaluation window of up to 30 seconds is available, the Area Under the ROC curve (AUROC) increases to 0.8306, indicating the model is effectively leveraging more information at a cost of slightly delayed decisions. This work opens new areas of research on how to incorporate the automatic weighing of human and machine expertise into human-AI pairing setups, which need to react dynamically to nonstationary expertise distribution between the human and AI players (e.g. when the experts need to be replaced, or the task at hand changes rapidly). Along with this paper, we offer the eye tracking data used in this study collected from 6 experts and 53 non-experts solving iris PAD visual task.
Related papers
- Human Scanpath Prediction in Target-Present Visual Search with Semantic-Foveal Bayesian Attention [49.99728312519117]
SemBA-FAST is a top-down framework designed for predicting human visual attention in target-present visual search.<n>We evaluate SemBA-FAST on the COCO-Search18 benchmark dataset, comparing its performance against other scanpath prediction models.<n>These findings provide valuable insights into the capabilities of semantic-foveal probabilistic frameworks for human-like attention modelling.
arXiv Detail & Related papers (2025-07-24T15:19:23Z) - Anomaly Detection Using Computer Vision: A Comparative Analysis of Class Distinction and Performance Metrics [0.0]
The study focuses on class distinction and performance evaluation, combining OpenCV with deep learning techniques.<n>The system effectively distinguishes among three classes: authorized personnel (admin), intruders, and non-human entities.
arXiv Detail & Related papers (2025-03-24T19:36:47Z) - AIDE: An Automatic Data Engine for Object Detection in Autonomous Driving [68.73885845181242]
We propose an Automatic Data Engine (AIDE) that automatically identifies issues, efficiently curates data, improves the model through auto-labeling, and verifies the model through generation of diverse scenarios.
We further establish a benchmark for open-world detection on AV datasets to comprehensively evaluate various learning paradigms, demonstrating our method's superior performance at a reduced cost.
arXiv Detail & Related papers (2024-03-26T04:27:56Z) - CLERA: A Unified Model for Joint Cognitive Load and Eye Region Analysis
in the Wild [18.79132232751083]
Real-time analysis of the dynamics of the eye region allows us to monitor humans' visual attention allocation and estimate their mental state.
We propose CLERA, which achieves precise keypoint detection andtemporal tracking in a joint-learning framework.
We also introduce a large-scale dataset of 30k human faces with joint pupil, eye-openness, and landmark annotation.
arXiv Detail & Related papers (2023-06-26T21:20:23Z) - Tackling Face Verification Edge Cases: In-Depth Analysis and
Human-Machine Fusion Approach [5.574995936464475]
This paper investigates the effect of a combination of machine and human operators in the face verification task.
We conduct a study with 60 participants on selected tasks with humans and provide an extensive analysis.
We demonstrate that combining machine and human decisions can further improve the performance of state-of-the-art face verification systems.
arXiv Detail & Related papers (2023-04-17T10:29:26Z) - Learning Transferable Pedestrian Representation from Multimodal
Information Supervision [174.5150760804929]
VAL-PAT is a novel framework that learns transferable representations to enhance various pedestrian analysis tasks with multimodal information.
We first perform pre-training on LUPerson-TA dataset, where each image contains text and attribute annotations.
We then transfer the learned representations to various downstream tasks, including person reID, person attribute recognition and text-based person search.
arXiv Detail & Related papers (2023-04-12T01:20:58Z) - An Efficient Point of Gaze Estimator for Low-Resolution Imaging Systems
Using Extracted Ocular Features Based Neural Architecture [2.8728982844941187]
This paper introduces a neural network based architecture to predict users' gaze at 9 positions displayed in the 11.31deg visual range on the screen.
The eye tracking system can be incorporated by physically disabled individuals, fitted best for those who have eyes as only a limited set of communication.
arXiv Detail & Related papers (2021-06-09T14:35:55Z) - Learnable Online Graph Representations for 3D Multi-Object Tracking [156.58876381318402]
We propose a unified and learning based approach to the 3D MOT problem.
We employ a Neural Message Passing network for data association that is fully trainable.
We show the merit of the proposed approach on the publicly available nuScenes dataset by achieving state-of-the-art performance of 65.6% AMOTA and 58% fewer ID-switches.
arXiv Detail & Related papers (2021-04-23T17:59:28Z) - TRiPOD: Human Trajectory and Pose Dynamics Forecasting in the Wild [77.59069361196404]
TRiPOD is a novel method for predicting body dynamics based on graph attentional networks.
To incorporate a real-world challenge, we learn an indicator representing whether an estimated body joint is visible/invisible at each frame.
Our evaluation shows that TRiPOD outperforms all prior work and state-of-the-art specifically designed for each of the trajectory and pose forecasting tasks.
arXiv Detail & Related papers (2021-04-08T20:01:00Z) - Automated analysis of eye-tracker-based human-human interaction studies [2.433293618209319]
We investigate which state-of-the-art computer vision algorithms may be used to automate the post-analysis of mobile eye-tracking data.
For the case study in this paper, we focus on mobile eye-tracker recordings made during human-human face-to-face interactions.
We show that the use of this single-pipeline framework provides robust results, which are both more accurate and faster than previous work in the field.
arXiv Detail & Related papers (2020-07-09T10:00:03Z) - AutoOD: Automated Outlier Detection via Curiosity-guided Search and
Self-imitation Learning [72.99415402575886]
Outlier detection is an important data mining task with numerous practical applications.
We propose AutoOD, an automated outlier detection framework, which aims to search for an optimal neural network model.
Experimental results on various real-world benchmark datasets demonstrate that the deep model identified by AutoOD achieves the best performance.
arXiv Detail & Related papers (2020-06-19T18:57:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.