A Study on Visual Perception of Light Field Content
- URL: http://arxiv.org/abs/2008.03195v1
- Date: Fri, 7 Aug 2020 14:23:27 GMT
- Title: A Study on Visual Perception of Light Field Content
- Authors: Ailbhe Gill, Emin Zerman, Cagri Ozcinar, Aljosa Smolic
- Abstract summary: We present a visual attention study on light field content.
We conducted perception experiments displaying them to users in various ways.
Our analysis highlights characteristics of user behaviour in light field imaging applications.
- Score: 19.397619552417986
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The effective design of visual computing systems depends heavily on the
anticipation of visual attention, or saliency. While visual attention is well
investigated for conventional 2D images and video, it is nevertheless a very
active research area for emerging immersive media. In particular, visual
attention of light fields (light rays of a scene captured by a grid of cameras
or micro lenses) has only recently become a focus of research. As they may be
rendered and consumed in various ways, a primary challenge that arises is the
definition of what visual perception of light field content should be. In this
work, we present a visual attention study on light field content. We conducted
perception experiments displaying them to users in various ways and collected
corresponding visual attention data. Our analysis highlights characteristics of
user behaviour in light field imaging applications. The light field data set
and attention data are provided with this paper.
Related papers
- Low-Light Enhancement Effect on Classification and Detection: An Empirical Study [48.6762437869172]
We evaluate the impact of Low-Light Image Enhancement (LLIE) methods on high-level vision tasks.
Our findings suggest a disconnect between image enhancement for human visual perception and for machine analysis.
This insight is crucial for the development of LLIE techniques that align with the needs of both human and machine vision.
arXiv Detail & Related papers (2024-09-22T14:21:31Z) - Semantic-Based Active Perception for Humanoid Visual Tasks with Foveal Sensors [49.99728312519117]
The aim of this work is to establish how accurately a recent semantic-based active perception model is able to complete visual tasks that are regularly performed by humans.
This model exploits the ability of current object detectors to localize and classify a large number of object classes and to update a semantic description of a scene across multiple fixations.
In the task of scene exploration, the semantic-based method demonstrates superior performance compared to the traditional saliency-based model.
arXiv Detail & Related papers (2024-04-16T18:15:57Z) - Stochastic Light Field Holography [35.73147050231529]
The Visual Turing Test is the ultimate goal to evaluate the realism of holographic displays.
Previous studies have focused on addressing challenges such as limited 'etendue and image quality over a large focal volume.
We tackle this problem with a novel hologram generation algorithm motivated by matching the projection operators of incoherent Light Field.
arXiv Detail & Related papers (2023-07-12T16:20:08Z) - Decoding Attention from Gaze: A Benchmark Dataset and End-to-End Models [6.642042615005632]
Eye-tracking has potential to provide rich behavioral data about human cognition in ecologically valid environments.
This paper studies using computer vision tools for "attention decoding", the task of assessing the locus of a participant's overt visual attention over time.
arXiv Detail & Related papers (2022-11-20T12:24:57Z) - Peripheral Vision Transformer [52.55309200601883]
We take a biologically inspired approach and explore to model peripheral vision in deep neural networks for visual recognition.
We propose to incorporate peripheral position encoding to the multi-head self-attention layers to let the network learn to partition the visual field into diverse peripheral regions given training data.
We evaluate the proposed network, dubbed PerViT, on the large-scale ImageNet dataset and systematically investigate the inner workings of the model for machine perception.
arXiv Detail & Related papers (2022-06-14T12:47:47Z) - A Dataset for Provident Vehicle Detection at Night [3.1969855247377827]
We study the problem of how to map this intuitive human behavior to computer vision algorithms to detect oncoming vehicles at night.
We present an extensive open-source dataset containing 59746 annotated grayscale images out of 346 different scenes in a rural environment at night.
We discuss the characteristics of the dataset and the challenges in objectively describing visual cues such as light reflections.
arXiv Detail & Related papers (2021-05-27T15:31:33Z) - Onfocus Detection: Identifying Individual-Camera Eye Contact from
Unconstrained Images [81.64699115587167]
Onfocus detection aims at identifying whether the focus of the individual captured by a camera is on the camera or not.
We build a large-scale onfocus detection dataset, named as the OnFocus Detection In the Wild (OFDIW)
We propose a novel end-to-end deep model, i.e., the eye-context interaction inferring network (ECIIN) for onfocus detection.
arXiv Detail & Related papers (2021-03-29T03:29:09Z) - What Can You Learn from Your Muscles? Learning Visual Representation
from Human Interactions [50.435861435121915]
We use human interaction and attention cues to investigate whether we can learn better representations compared to visual-only representations.
Our experiments show that our "muscly-supervised" representation outperforms a visual-only state-of-the-art method MoCo.
arXiv Detail & Related papers (2020-10-16T17:46:53Z) - Deep learning for scene recognition from visual data: a survey [2.580765958706854]
This work aims to be a review of the state-of-the-art in scene recognition with deep learning models from visual data.
Scene recognition is still an emerging field in computer vision, which has been addressed from a single image and dynamic image perspective.
arXiv Detail & Related papers (2020-07-03T16:53:18Z) - Multi-Modal Music Information Retrieval: Augmenting Audio-Analysis with
Visual Computing for Improved Music Video Analysis [91.3755431537592]
This thesis combines audio-analysis with computer vision to approach Music Information Retrieval (MIR) tasks from a multi-modal perspective.
The main hypothesis of this work is based on the observation that certain expressive categories such as genre or theme can be recognized on the basis of the visual content alone.
The experiments are conducted for three MIR tasks Artist Identification, Music Genre Classification and Cross-Genre Classification.
arXiv Detail & Related papers (2020-02-01T17:57:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.