Exploring the Effectiveness of Video Perceptual Representation in Blind
Video Quality Assessment
- URL: http://arxiv.org/abs/2207.03723v1
- Date: Fri, 8 Jul 2022 07:30:51 GMT
- Title: Exploring the Effectiveness of Video Perceptual Representation in Blind
Video Quality Assessment
- Authors: Liang Liao, Kangmin Xu, Haoning Wu, Chaofeng Chen, Wenxiu Sun, Qiong
Yan, Weisi Lin
- Abstract summary: We propose a temporal perceptual quality index (TPQI) to measure the temporal distortion by describing the graphic morphology of the representation.
Experiments show that TPQI is an effective way of predicting subjective temporal quality.
- Score: 55.65173181828863
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: With the rapid growth of in-the-wild videos taken by non-specialists, blind
video quality assessment (VQA) has become a challenging and demanding problem.
Although lots of efforts have been made to solve this problem, it remains
unclear how the human visual system (HVS) relates to the temporal quality of
videos. Meanwhile, recent work has found that the frames of natural video
transformed into the perceptual domain of the HVS tend to form a straight
trajectory of the representations. With the obtained insight that distortion
impairs the perceived video quality and results in a curved trajectory of the
perceptual representation, we propose a temporal perceptual quality index
(TPQI) to measure the temporal distortion by describing the graphic morphology
of the representation. Specifically, we first extract the video perceptual
representations from the lateral geniculate nucleus (LGN) and primary visual
area (V1) of the HVS, and then measure the straightness and compactness of
their trajectories to quantify the degradation in naturalness and content
continuity of video. Experiments show that the perceptual representation in the
HVS is an effective way of predicting subjective temporal quality, and thus
TPQI can, for the first time, achieve comparable performance to the spatial
quality metric and be even more effective in assessing videos with large
temporal variations. We further demonstrate that by combining with NIQE, a
spatial quality metric, TPQI can achieve top performance over popular
in-the-wild video datasets. More importantly, TPQI does not require any
additional information beyond the video being evaluated and thus can be applied
to any datasets without parameter tuning. Source code is available at
https://github.com/UoLMM/TPQI-VQA.
Related papers
- CLIPVQA:Video Quality Assessment via CLIP [56.94085651315878]
We propose an efficient CLIP-based Transformer method for the VQA problem ( CLIPVQA)
The proposed CLIPVQA achieves new state-of-the-art VQA performance and up to 37% better generalizability than existing benchmark VQA methods.
arXiv Detail & Related papers (2024-07-06T02:32:28Z) - On the Content Bias in Fréchet Video Distance [42.717821654744796]
Fr'echet Video Distance (FVD) is a prominent metric for evaluating video generation models.
In this paper, we aim to explore the extent of FVD's bias toward per-frame quality over temporal realism.
We show that FVD with features extracted from the recent large-scale self-supervised video models is less biased toward image quality.
arXiv Detail & Related papers (2024-04-18T17:59:58Z) - Capturing Co-existing Distortions in User-Generated Content for
No-reference Video Quality Assessment [9.883856205077022]
Video Quality Assessment (VQA) aims to predict the perceptual quality of a video.
VQA faces two under-estimated challenges unresolved in User Generated Content (UGC) videos.
We propose textitVisual Quality Transformer (VQT) to extract quality-related sparse features more efficiently.
arXiv Detail & Related papers (2023-07-31T16:29:29Z) - Evaluating Point Cloud from Moving Camera Videos: A No-Reference Metric [58.309735075960745]
This paper explores the way of dealing with point cloud quality assessment (PCQA) tasks via video quality assessment (VQA) methods.
We generate the captured videos by rotating the camera around the point clouds through several circular pathways.
We extract both spatial and temporal quality-aware features from the selected key frames and the video clips through using trainable 2D-CNN and pre-trained 3D-CNN models.
arXiv Detail & Related papers (2022-08-30T08:59:41Z) - DisCoVQA: Temporal Distortion-Content Transformers for Video Quality
Assessment [56.42140467085586]
Some temporal variations are causing temporal distortions and lead to extra quality degradations.
Human visual system often has different attention to frames with different contents.
We propose a novel and effective transformer-based VQA method to tackle these two issues.
arXiv Detail & Related papers (2022-06-20T15:31:27Z) - Blind VQA on 360{\deg} Video via Progressively Learning from Pixels,
Frames and Video [66.57045901742922]
Blind visual quality assessment (BVQA) on 360textdegree video plays a key role in optimizing immersive multimedia systems.
In this paper, we take into account the progressive paradigm of human perception towards spherical video quality.
We propose a novel BVQA approach (namely ProVQA) for 360textdegree video via progressively learning from pixels, frames and video.
arXiv Detail & Related papers (2021-11-18T03:45:13Z) - FOVQA: Blind Foveated Video Quality Assessment [1.4127304025810108]
We develop a no-reference (NR) foveated video quality assessment model, called FOVQA.
It is based on new models of space-variant natural scene statistics (NSS) and natural video statistics (NVS)
FOVQA achieves state-of-the-art (SOTA) performance on the new 2D LIVE-FBT-FCVR database.
arXiv Detail & Related papers (2021-06-24T21:38:22Z) - ST-GREED: Space-Time Generalized Entropic Differences for Frame Rate
Dependent Video Quality Prediction [63.749184706461826]
We study how perceptual quality is affected by frame rate, and how frame rate and compression combine to affect perceived quality.
We devise an objective VQA model called Space-Time GeneRalized Entropic Difference (GREED) which analyzes the statistics of spatial and temporal band-pass video coefficients.
GREED achieves state-of-the-art performance on the LIVE-YT-HFR Database when compared with existing VQA models.
arXiv Detail & Related papers (2020-10-26T16:54:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.