Deep semantic gaze embedding and scanpath comparison for expertise
classification during OPT viewing
- URL: http://arxiv.org/abs/2003.13987v1
- Date: Tue, 31 Mar 2020 07:00:59 GMT
- Title: Deep semantic gaze embedding and scanpath comparison for expertise
classification during OPT viewing
- Authors: Nora Castner, Thomas K\"ubler, Katharina Scheiter, Juilane Richter,
Th\'er\'ese Eder, Fabian H\"uttig, Constanze Keutel, Enkelejda Kasneci
- Abstract summary: We present a novel approach to gaze scanpath comparison that incorporates convolutional neural networks (CNN)
Our approach was capable of distinguishing experts from novices with 93% accuracy while incorporating the image semantics.
- Score: 6.700983301090583
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Modeling eye movement indicative of expertise behavior is decisive in user
evaluation. However, it is indisputable that task semantics affect gaze
behavior. We present a novel approach to gaze scanpath comparison that
incorporates convolutional neural networks (CNN) to process scene information
at the fixation level. Image patches linked to respective fixations are used as
input for a CNN and the resulting feature vectors provide the temporal and
spatial gaze information necessary for scanpath similarity comparison.We
evaluated our proposed approach on gaze data from expert and novice dentists
interpreting dental radiographs using a local alignment similarity score. Our
approach was capable of distinguishing experts from novices with 93% accuracy
while incorporating the image semantics. Moreover, our scanpath comparison
using image patch features has the potential to incorporate task semantics from
a variety of tasks
Related papers
- Graph Self-Supervised Learning for Endoscopic Image Matching [1.8275108630751844]
We propose a novel self-supervised approach that combines Convolutional Neural Networks for capturing local visual appearance and attention-based Graph Neural Networks for modeling spatial relationships between key-points.
Our approach is trained in a fully self-supervised scheme without the need for labeled data.
Our approach outperforms state-of-the-art handcrafted and deep learning-based methods, demonstrating exceptional performance in terms of precision rate (1) and matching score (99.3%)
arXiv Detail & Related papers (2023-06-19T19:53:41Z) - Location-Aware Self-Supervised Transformers [74.76585889813207]
We propose to pretrain networks for semantic segmentation by predicting the relative location of image parts.
We control the difficulty of the task by masking a subset of the reference patch features visible to those of the query.
Our experiments show that this location-aware pretraining leads to representations that transfer competitively to several challenging semantic segmentation benchmarks.
arXiv Detail & Related papers (2022-12-05T16:24:29Z) - An Inter-observer consistent deep adversarial training for visual
scanpath prediction [66.46953851227454]
We propose an inter-observer consistent adversarial training approach for scanpath prediction through a lightweight deep neural network.
We show the competitiveness of our approach in regard to state-of-the-art methods.
arXiv Detail & Related papers (2022-11-14T13:22:29Z) - Two-Stream Graph Convolutional Network for Intra-oral Scanner Image
Segmentation [133.02190910009384]
We propose a two-stream graph convolutional network (i.e., TSGCN) to handle inter-view confusion between different raw attributes.
Our TSGCN significantly outperforms state-of-the-art methods in 3D tooth (surface) segmentation.
arXiv Detail & Related papers (2022-04-19T10:41:09Z) - Weakly supervised semantic segmentation of tomographic images in the
diagnosis of stroke [0.0]
This paper presents an automatic algorithm for the segmentation of areas affected by an acute stroke on the non-contrast computed tomography brain images.
The proposed algorithm is designed for learning in a weakly supervised scenario when some images are labeled accurately, and some images are labeled inaccurately.
arXiv Detail & Related papers (2021-09-04T15:24:38Z) - Scene Graph Embeddings Using Relative Similarity Supervision [4.137464623395376]
We employ a graph convolutional network to exploit structure in scene graphs and produce image embeddings useful for semantic image retrieval.
We propose a novel loss function that operates on pairs of similar and dissimilar images and imposes relative ordering between them in embedding space.
We demonstrate that this Ranking loss, coupled with an intuitive triple sampling strategy, leads to robust representations that outperform well-known contrastive losses on the retrieval task.
arXiv Detail & Related papers (2021-04-06T09:13:05Z) - Content-Based Detection of Temporal Metadata Manipulation [91.34308819261905]
We propose an end-to-end approach to verify whether the purported time of capture of an image is consistent with its content and geographic location.
The central idea is the use of supervised consistency verification, in which we predict the probability that the image content, capture time, and geographical location are consistent.
Our approach improves upon previous work on a large benchmark dataset, increasing the classification accuracy from 59.03% to 81.07%.
arXiv Detail & Related papers (2021-03-08T13:16:19Z) - Geography-Aware Self-Supervised Learning [79.4009241781968]
We show that due to their different characteristics, a non-trivial gap persists between contrastive and supervised learning on standard benchmarks.
We propose novel training methods that exploit the spatially aligned structure of remote sensing data.
Our experiments show that our proposed method closes the gap between contrastive and supervised learning on image classification, object detection and semantic segmentation for remote sensing.
arXiv Detail & Related papers (2020-11-19T17:29:13Z) - Graph Neural Networks for UnsupervisedDomain Adaptation of
Histopathological ImageAnalytics [22.04114134677181]
We present a novel method for the unsupervised domain adaptation for histological image analysis.
It is based on a backbone for embedding images into a feature space, and a graph neural layer for propa-gating the supervision signals of images with labels.
In experiments, our methodachieves state-of-the-art performance on four public datasets.
arXiv Detail & Related papers (2020-08-21T04:53:44Z) - Geometrically Mappable Image Features [85.81073893916414]
Vision-based localization of an agent in a map is an important problem in robotics and computer vision.
We propose a method that learns image features targeted for image-retrieval-based localization.
arXiv Detail & Related papers (2020-03-21T15:36:38Z) - Relevance Prediction from Eye-movements Using Semi-interpretable
Convolutional Neural Networks [9.007191808968242]
We propose an image-classification method to predict the perceived-relevance of text documents from eye-movements.
An eye-tracking study was conducted where participants read short news articles, and rated them as relevant or irrelevant for answering a trigger question.
We encode participants' eye-movement scanpaths as images, and then train a convolutional neural network classifier using these scanpath images.
arXiv Detail & Related papers (2020-01-15T07:02:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.