Target Features Affect Visual Search, A Study of Eye Fixations
- URL: http://arxiv.org/abs/2209.13771v1
- Date: Wed, 28 Sep 2022 01:53:16 GMT
- Title: Target Features Affect Visual Search, A Study of Eye Fixations
- Authors: Manoosh Samiei, James J. Clark
- Abstract summary: We investigate how the performance of human participants during visual search is affected by different parameters.
Our studies show that a bigger and more eccentric target is found faster with fewer number of fixations.
- Score: 2.7920304852537527
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Visual Search is referred to the task of finding a target object among a set
of distracting objects in a visual display. In this paper, based on an
independent analysis of the COCO-Search18 dataset, we investigate how the
performance of human participants during visual search is affected by different
parameters such as the size and eccentricity of the target object. We also
study the correlation between the error rate of participants and search
performance. Our studies show that a bigger and more eccentric target is found
faster with fewer number of fixations. Our code for the graphics are publicly
available at: \url{https://github.com/ManooshSamiei/COCOSearch18_Analysis}
Related papers
- Correlation of Object Detection Performance with Visual Saliency and Depth Estimation [0.09208007322096533]
This paper investigates the correlations between object detection accuracy and two fundamental visual tasks: depth prediction and visual saliency prediction.
Our analysis reveals significant variations in these correlations across object categories, with larger objects showing correlation values up to three times higher than smaller objects.
These findings suggest incorporating visual saliency features into object detection architectures could be more beneficial than depth information.
arXiv Detail & Related papers (2024-11-05T06:34:19Z) - OAT: Object-Level Attention Transformer for Gaze Scanpath Prediction [0.2796197251957245]
This paper introduces the Object-level Attention Transformer (OAT)
OAT predicts human scanpaths as they search for a target object within a cluttered scene of distractors.
We evaluate OAT on the Amazon book cover dataset and a new dataset for visual search that we collected.
arXiv Detail & Related papers (2024-07-18T09:33:17Z) - Semantic-Based Active Perception for Humanoid Visual Tasks with Foveal Sensors [49.99728312519117]
The aim of this work is to establish how accurately a recent semantic-based active perception model is able to complete visual tasks that are regularly performed by humans.
This model exploits the ability of current object detectors to localize and classify a large number of object classes and to update a semantic description of a scene across multiple fixations.
In the task of scene exploration, the semantic-based method demonstrates superior performance compared to the traditional saliency-based model.
arXiv Detail & Related papers (2024-04-16T18:15:57Z) - Visual Context-Aware Person Fall Detection [52.49277799455569]
We present a segmentation pipeline to semi-automatically separate individuals and objects in images.
Background objects such as beds, chairs, or wheelchairs can challenge fall detection systems, leading to false positive alarms.
We demonstrate that object-specific contextual transformations during training effectively mitigate this challenge.
arXiv Detail & Related papers (2024-04-11T19:06:36Z) - SeeBel: Seeing is Believing [0.9790236766474201]
We propose three visualizations that enable users to compare dataset statistics and AI performance for segmenting all images.
Our project tries to further increase the interpretability of the trained AI model for segmentation by visualizing its image attention weights.
We propose to conduct surveys on real users to study the efficacy of our visualization tool in computer vision and AI domain.
arXiv Detail & Related papers (2023-12-18T05:11:00Z) - Unified Visual Relationship Detection with Vision and Language Models [89.77838890788638]
This work focuses on training a single visual relationship detector predicting over the union of label spaces from multiple datasets.
We propose UniVRD, a novel bottom-up method for Unified Visual Relationship Detection by leveraging vision and language models.
Empirical results on both human-object interaction detection and scene-graph generation demonstrate the competitive performance of our model.
arXiv Detail & Related papers (2023-03-16T00:06:28Z) - Efficient Zero-shot Visual Search via Target and Context-aware
Transformer [5.652978777706897]
We propose a zero-shot deep learning architecture, TCT, that modulates self attention in the Vision Transformer with target and contextual relevant information.
We conduct visual search experiments on TCT and other competitive visual search models on three natural scene datasets with varying levels of difficulty.
TCT demonstrates human-like performance in terms of search efficiency and beats the SOTA models in challenging visual search tasks.
arXiv Detail & Related papers (2022-11-24T08:27:47Z) - Predicting Visual Attention and Distraction During Visual Search Using
Convolutional Neural Networks [2.7920304852537527]
We present two approaches to model visual attention and distraction of observers during visual search.
Our first approach adapts a light-weight free-viewing saliency model to predict eye fixation density maps of human observers over pixels of search images.
Our second approach is object-based and predicts the distractor and target objects during visual search.
arXiv Detail & Related papers (2022-10-27T00:39:43Z) - Learning Target Candidate Association to Keep Track of What Not to Track [100.80610986625693]
We propose to keep track of distractor objects in order to continue tracking the target.
To tackle the problem of lacking ground-truth correspondences between distractor objects in visual tracking, we propose a training strategy that combines partial annotations with self-supervision.
Our tracker sets a new state-of-the-art on six benchmarks, achieving an AUC score of 67.2% on LaSOT and a +6.1% absolute gain on the OxUvA long-term dataset.
arXiv Detail & Related papers (2021-03-30T17:58:02Z) - DRG: Dual Relation Graph for Human-Object Interaction Detection [65.50707710054141]
We tackle the challenging problem of human-object interaction (HOI) detection.
Existing methods either recognize the interaction of each human-object pair in isolation or perform joint inference based on complex appearance-based features.
In this paper, we leverage an abstract spatial-semantic representation to describe each human-object pair and aggregate the contextual information of the scene via a dual relation graph.
arXiv Detail & Related papers (2020-08-26T17:59:40Z) - ConsNet: Learning Consistency Graph for Zero-Shot Human-Object
Interaction Detection [101.56529337489417]
We consider the problem of Human-Object Interaction (HOI) Detection, which aims to locate and recognize HOI instances in the form of human, action, object> in images.
We argue that multi-level consistencies among objects, actions and interactions are strong cues for generating semantic representations of rare or previously unseen HOIs.
Our model takes visual features of candidate human-object pairs and word embeddings of HOI labels as inputs, maps them into visual-semantic joint embedding space and obtains detection results by measuring their similarities.
arXiv Detail & Related papers (2020-08-14T09:11:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.