Scanpath Prediction on Information Visualisations
- URL: http://arxiv.org/abs/2112.02340v1
- Date: Sat, 4 Dec 2021 13:59:52 GMT
- Title: Scanpath Prediction on Information Visualisations
- Authors: Yao Wang, Mihai B\^ace, and Andreas Bulling
- Abstract summary: We propose a model that learns to predict visual saliency and scanpaths on information visualisations.
We present in-depth analyses of gaze behaviour for different information visualisation elements on the popular MASSVIS dataset.
- Score: 19.591855190022667
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose Unified Model of Saliency and Scanpaths (UMSS) -- a model that
learns to predict visual saliency and scanpaths (i.e. sequences of eye
fixations) on information visualisations. Although scanpaths provide rich
information about the importance of different visualisation elements during the
visual exploration process, prior work has been limited to predicting
aggregated attention statistics, such as visual saliency. We present in-depth
analyses of gaze behaviour for different information visualisation elements
(e.g. Title, Label, Data) on the popular MASSVIS dataset. We show that while,
overall, gaze patterns are surprisingly consistent across visualisations and
viewers, there are also structural differences in gaze dynamics for different
elements. Informed by our analyses, UMSS first predicts multi-duration
element-level saliency maps, then probabilistically samples scanpaths from
them. Extensive experiments on MASSVIS show that our method consistently
outperforms state-of-the-art methods with respect to several, widely used
scanpath and saliency evaluation metrics. Our method achieves a relative
improvement in sequence score of 11.5% for scanpath prediction, and a relative
improvement in Pearson correlation coefficient of up to 23.6% for saliency
prediction. These results are auspicious and point towards richer user models
and simulations of visual attention on visualisations without the need for any
eye tracking equipment.
Related papers
- Diffusion Models as Data Mining Tools [87.77999285241219]
This paper demonstrates how to use generative models trained for image synthesis as tools for visual data mining.
We show that after finetuning conditional diffusion models to synthesize images from a specific dataset, we can use these models to define a typicality measure.
This measure assesses how typical visual elements are for different data labels, such as geographic location, time stamps, semantic labels, or even the presence of a disease.
arXiv Detail & Related papers (2024-07-20T17:14:31Z) - Beyond Average: Individualized Visual Scanpath Prediction [20.384132849805003]
individualized scanpath prediction (ISP) aims to accurately predict how different individuals shift their attention in diverse visual tasks.
ISP features an observer encoder to characterize and integrate an observer's unique attention traits, an observer-centric feature integration approach, and an adaptive fixation prioritization mechanism.
Our method is generally applicable to different datasets, model architectures, and visual tasks, offering a comprehensive tool for transforming general scanpath models into individualized ones.
arXiv Detail & Related papers (2024-04-18T14:51:42Z) - Semantic-Based Active Perception for Humanoid Visual Tasks with Foveal Sensors [49.99728312519117]
The aim of this work is to establish how accurately a recent semantic-based active perception model is able to complete visual tasks that are regularly performed by humans.
This model exploits the ability of current object detectors to localize and classify a large number of object classes and to update a semantic description of a scene across multiple fixations.
In the task of scene exploration, the semantic-based method demonstrates superior performance compared to the traditional saliency-based model.
arXiv Detail & Related papers (2024-04-16T18:15:57Z) - TempSAL -- Uncovering Temporal Information for Deep Saliency Prediction [64.63645677568384]
We introduce a novel saliency prediction model that learns to output saliency maps in sequential time intervals.
Our approach locally modulates the saliency predictions by combining the learned temporal maps.
Our code will be publicly available on GitHub.
arXiv Detail & Related papers (2023-01-05T22:10:16Z) - Active Gaze Control for Foveal Scene Exploration [124.11737060344052]
We propose a methodology to emulate how humans and robots with foveal cameras would explore a scene.
The proposed method achieves an increase in detection F1-score of 2-3 percentage points for the same number of gaze shifts.
arXiv Detail & Related papers (2022-08-24T14:59:28Z) - A Graph-Enhanced Click Model for Web Search [67.27218481132185]
We propose a novel graph-enhanced click model (GraphCM) for web search.
We exploit both intra-session and inter-session information for the sparsity and cold-start problems.
arXiv Detail & Related papers (2022-06-17T08:32:43Z) - A Simple and efficient deep Scanpath Prediction [6.294759639481189]
We explore the efficiency of using common deep learning architectures, in a simple fully convolutional regressive manner.
We experiment how well these models can predict the scanpaths on 2 datasets.
We also compare the different leveraged backbone architectures based on their performances on the experiment to deduce which ones are the most suitable for the task.
arXiv Detail & Related papers (2021-12-08T22:43:45Z) - VizAI : Selecting Accurate Visualizations of Numerical Data [2.6039035727217907]
VizAI is a generative-discriminative framework that first generates various statistical properties of the data.
It is linked to a discriminative model that selects the visualization that best matches the true statistics of the data being visualized.
VizAI can easily be trained with minimal supervision and adapts to settings with varying degrees of supervision easily.
arXiv Detail & Related papers (2021-11-07T22:05:44Z) - SALYPATH: A Deep-Based Architecture for visual attention prediction [5.068678962285629]
Visual attention is useful for many computer vision applications such as image compression, recognition, and captioning.
We propose an end-to-end deep-based method, so-called SALYPATH, that efficiently predicts the scanpath of an image through features of a saliency model.
The idea is predict the scanpath by exploiting the capacity of a deep-based model to predict the saliency.
arXiv Detail & Related papers (2021-06-29T08:53:51Z) - Adversarial Feature Augmentation and Normalization for Visual
Recognition [109.6834687220478]
Recent advances in computer vision take advantage of adversarial data augmentation to ameliorate the generalization ability of classification models.
Here, we present an effective and efficient alternative that advocates adversarial augmentation on intermediate feature embeddings.
We validate the proposed approach across diverse visual recognition tasks with representative backbone networks.
arXiv Detail & Related papers (2021-03-22T20:36:34Z) - Classifying Eye-Tracking Data Using Saliency Maps [8.524684315458245]
This paper proposes a visual saliency based novel feature extraction method for automatic and quantitative classification of eye-tracking data.
Comparing the saliency amplitudes, similarity and dissimilarity of saliency maps with the corresponding eye fixations maps gives an extra dimension of information which is effectively utilized to generate discriminative features to classify the eye-tracking data.
arXiv Detail & Related papers (2020-10-24T15:18:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.