ST360IQ: No-Reference Omnidirectional Image Quality Assessment with
Spherical Vision Transformers
- URL: http://arxiv.org/abs/2303.06907v1
- Date: Mon, 13 Mar 2023 07:48:46 GMT
- Title: ST360IQ: No-Reference Omnidirectional Image Quality Assessment with
Spherical Vision Transformers
- Authors: Nafiseh Jabbari Tofighi, Mohamed Hedi Elfkir, Nevrez Imamoglu, Cagri
Ozcinar, Erkut Erdem, Aykut Erdem
- Abstract summary: We present a method for no-reference 360 image quality assessment.
Our approach predicts the quality of an omnidirectional image correlated with the human-perceived image quality.
- Score: 17.48330099000856
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Omnidirectional images, aka 360 images, can deliver immersive and interactive
visual experiences. As their popularity has increased dramatically in recent
years, evaluating the quality of 360 images has become a problem of interest
since it provides insights for capturing, transmitting, and consuming this new
media. However, directly adapting quality assessment methods proposed for
standard natural images for omnidirectional data poses certain challenges.
These models need to deal with very high-resolution data and implicit
distortions due to the spherical form of the images. In this study, we present
a method for no-reference 360 image quality assessment. Our proposed ST360IQ
model extracts tangent viewports from the salient parts of the input
omnidirectional image and employs a vision-transformers based module processing
saliency selective patches/tokens that estimates a quality score from each
viewport. Then, it aggregates these scores to give a final quality score. Our
experiments on two benchmark datasets, namely OIQA and CVIQ datasets,
demonstrate that as compared to the state-of-the-art, our approach predicts the
quality of an omnidirectional image correlated with the human-perceived image
quality. The code has been available on
https://github.com/Nafiseh-Tofighi/ST360IQ
Related papers
- Perceptual Depth Quality Assessment of Stereoscopic Omnidirectional Images [10.382801621282228]
We develop an objective quality assessment model named depth quality index (DQI) for efficient no-reference (NR) depth quality assessment of stereoscopic omnidirectional images.
Motivated by the perceptual characteristics of the human visual system (HVS), the proposed DQI is built upon multi-color-channel, adaptive viewport selection, and interocular discrepancy features.
arXiv Detail & Related papers (2024-08-19T16:28:05Z) - AIGCOIQA2024: Perceptual Quality Assessment of AI Generated Omnidirectional Images [70.42666704072964]
We establish a large-scale AI generated omnidirectional image IQA database named AIGCOIQA2024.
A subjective IQA experiment is conducted to assess human visual preferences from three perspectives.
We conduct a benchmark experiment to evaluate the performance of state-of-the-art IQA models on our database.
arXiv Detail & Related papers (2024-04-01T10:08:23Z) - Perceptual Quality Assessment of 360$^\circ$ Images Based on Generative
Scanpath Representation [40.00063797833765]
We introduce a unique generative scanpath representation (GSR) for effective quality inference of 360$circ$ images.
GSR aggregates varied perceptual experiences of multi-hypothesis users under a predefined viewing condition.
We then propose an efficient OIQA computational framework by learning the quality maps GSR.
arXiv Detail & Related papers (2023-09-07T04:10:30Z) - Assessor360: Multi-sequence Network for Blind Omnidirectional Image
Quality Assessment [50.82681686110528]
Blind Omnidirectional Image Quality Assessment (BOIQA) aims to objectively assess the human perceptual quality of omnidirectional images (ODIs)
The quality assessment of ODIs is severely hampered by the fact that the existing BOIQA pipeline lacks the modeling of the observer's browsing process.
We propose a novel multi-sequence network for BOIQA called Assessor360, which is derived from the realistic multi-assessor ODI quality assessment procedure.
arXiv Detail & Related papers (2023-05-18T13:55:28Z) - End-to-end deep multi-score model for No-reference stereoscopic image
quality assessment [6.254148286968409]
We use a deep multi-score Convolutional Neural Network (CNN) to estimate stereoscopic image quality without reference.
Our model has been trained to perform four tasks: First, predict the left view's quality. Second, predict the quality of the left view. Third and fourth, predict the quality of the stereo view and global quality, respectively, with the global score serving as the ultimate quality.
arXiv Detail & Related papers (2022-11-02T16:45:35Z) - Perceptual Quality Assessment of Omnidirectional Images [81.76416696753947]
We first establish an omnidirectional IQA (OIQA) database, which includes 16 source images and 320 distorted images degraded by 4 commonly encountered distortion types.
Then a subjective quality evaluation study is conducted on the OIQA database in the VR environment.
The original and distorted omnidirectional images, subjective quality ratings, and the head and eye movement data together constitute the OIQA database.
arXiv Detail & Related papers (2022-07-06T13:40:38Z) - Conformer and Blind Noisy Students for Improved Image Quality Assessment [80.57006406834466]
Learning-based approaches for perceptual image quality assessment (IQA) usually require both the distorted and reference image for measuring the perceptual quality accurately.
In this work, we explore the performance of transformer-based full-reference IQA models.
We also propose a method for IQA based on semi-supervised knowledge distillation from full-reference teacher models into blind student models.
arXiv Detail & Related papers (2022-04-27T10:21:08Z) - Blind VQA on 360{\deg} Video via Progressively Learning from Pixels,
Frames and Video [66.57045901742922]
Blind visual quality assessment (BVQA) on 360textdegree video plays a key role in optimizing immersive multimedia systems.
In this paper, we take into account the progressive paradigm of human perception towards spherical video quality.
We propose a novel BVQA approach (namely ProVQA) for 360textdegree video via progressively learning from pixels, frames and video.
arXiv Detail & Related papers (2021-11-18T03:45:13Z) - Perceptual Image Quality Assessment with Transformers [4.005576542371173]
We propose an image quality transformer (IQT) that successfully applies a transformer architecture to a perceptual full-reference image quality assessment task.
We extract the perceptual feature representations from each of input images using a convolutional neural network backbone.
The proposed IQT was ranked first among 13 participants in the NTIRE 2021 perceptual image quality assessment challenge.
arXiv Detail & Related papers (2021-04-30T02:45:29Z) - No-Reference Quality Assessment for 360-degree Images by Analysis of
Multi-frequency Information and Local-global Naturalness [26.614657212889398]
360-degree/omnidirectional images (OIs) have achieved remarkable attentions due to the increasing applications of virtual reality (VR)
We propose a novel and effective no-reference omnidirectional image quality assessment (NR OIQA) algorithm by Multi-Frequency Information and Local-Global Naturalness (MFILGN)
Experimental results on two publicly available OIQA databases demonstrate that our proposed MFILGN outperforms state-of-the-art approaches.
arXiv Detail & Related papers (2021-02-22T22:52:35Z) - Visual Question Answering on 360{\deg} Images [96.00046925811515]
VQA 360 is a novel task of visual question answering on 360 images.
We collect the first VQA 360 dataset, containing around 17,000 real-world image-question-answer triplets for a variety of question types.
arXiv Detail & Related papers (2020-01-10T08:18:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.