StableVQA: A Deep No-Reference Quality Assessment Model for Video
Stability
- URL: http://arxiv.org/abs/2308.04904v3
- Date: Fri, 27 Oct 2023 09:13:38 GMT
- Title: StableVQA: A Deep No-Reference Quality Assessment Model for Video
Stability
- Authors: Tengchuan Kou, Xiaohong Liu, Wei Sun, Jun Jia, Xiongkuo Min, Guangtao
Zhai, Ning Liu
- Abstract summary: Video shakiness is an unpleasant distortion of User Generated Content (UGC) videos, which is usually caused by the unstable hold of cameras.
We build a new database named 1,952 diversely-shaky videos, where each video has a Mean Opinion Score (MOS) on the degree of video stability rated by 34 subjects.
We elaborately design a novel VQA-S model named StableVQA, which consists of three feature extractors to acquire the optical flow, semantic, and blur features respectively, and a regression layer to predict the final stability score.
- Score: 56.462032266188785
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Video shakiness is an unpleasant distortion of User Generated Content (UGC)
videos, which is usually caused by the unstable hold of cameras. In recent
years, many video stabilization algorithms have been proposed, yet no specific
and accurate metric enables comprehensively evaluating the stability of videos.
Indeed, most existing quality assessment models evaluate video quality as a
whole without specifically taking the subjective experience of video stability
into consideration. Therefore, these models cannot measure the video stability
explicitly and precisely when severe shakes are present. In addition, there is
no large-scale video database in public that includes various degrees of shaky
videos with the corresponding subjective scores available, which hinders the
development of Video Quality Assessment for Stability (VQA-S). To this end, we
build a new database named StableDB that contains 1,952 diversely-shaky UGC
videos, where each video has a Mean Opinion Score (MOS) on the degree of video
stability rated by 34 subjects. Moreover, we elaborately design a novel VQA-S
model named StableVQA, which consists of three feature extractors to acquire
the optical flow, semantic, and blur features respectively, and a regression
layer to predict the final stability score. Extensive experiments demonstrate
that the StableVQA achieves a higher correlation with subjective opinions than
the existing VQA-S models and generic VQA models. The database and codes are
available at https://github.com/QMME/StableVQA.
Related papers
- Benchmarking AIGC Video Quality Assessment: A Dataset and Unified Model [54.69882562863726]
We try to systemically investigate the AIGC-VQA problem from both subjective and objective quality assessment perspectives.
We evaluate the perceptual quality of AIGC videos from three dimensions: spatial quality, temporal quality, and text-to-video alignment.
We propose a Unify Generated Video Quality assessment (UGVQ) model to comprehensively and accurately evaluate the quality of AIGC videos.
arXiv Detail & Related papers (2024-07-31T07:54:26Z) - Enhancing Blind Video Quality Assessment with Rich Quality-aware Features [79.18772373737724]
We present a simple but effective method to enhance blind video quality assessment (BVQA) models for social media videos.
We explore rich quality-aware features from pre-trained blind image quality assessment (BIQA) and BVQA models as auxiliary features.
Experimental results demonstrate that the proposed model achieves the best performance on three public social media VQA datasets.
arXiv Detail & Related papers (2024-05-14T16:32:11Z) - Modular Blind Video Quality Assessment [33.657933680973194]
Blind video quality assessment (BVQA) plays a pivotal role in evaluating and improving the viewing experience of end-users across a wide range of video-based platforms and services.
In this paper, we propose a modular BVQA model and a method of training it to improve its modularity.
arXiv Detail & Related papers (2024-02-29T15:44:00Z) - KVQ: Kwai Video Quality Assessment for Short-form Videos [24.5291786508361]
We establish the first large-scale Kaleidoscope short Video database for Quality assessment, KVQ, which comprises 600 user-uploaded short videos and 3600 processed videos.
We propose the first short-form video quality evaluator, i.e., KSVQE, which enables the quality evaluator to identify the quality-determined semantics with the content understanding of large vision language models.
arXiv Detail & Related papers (2024-02-11T14:37:54Z) - Analysis of Video Quality Datasets via Design of Minimalistic Video Quality Models [71.06007696593704]
Blind quality assessment (BVQA) plays an indispensable role in monitoring and improving the end-users' viewing experience in real-world video-enabled media applications.
As an experimental field, the improvements of BVQA models have been measured primarily on a few human-rated VQA datasets.
We conduct a first-of-its-kind computational analysis of VQA datasets via minimalistic BVQA models.
arXiv Detail & Related papers (2023-07-26T06:38:33Z) - Towards Explainable In-the-Wild Video Quality Assessment: A Database and
a Language-Prompted Approach [52.07084862209754]
We collect over two million opinions on 4,543 in-the-wild videos on 13 dimensions of quality-related factors.
Specifically, we ask the subjects to label among a positive, a negative, and a neutral choice for each dimension.
These explanation-level opinions allow us to measure the relationships between specific quality factors and abstract subjective quality ratings.
arXiv Detail & Related papers (2023-05-22T05:20:23Z) - UGC-VQA: Benchmarking Blind Video Quality Assessment for User Generated
Content [59.13821614689478]
Blind quality prediction of in-the-wild videos is quite challenging, since the quality degradations of content are unpredictable, complicated, and often commingled.
Here we contribute to advancing the problem by conducting a comprehensive evaluation of leading VQA models.
By employing a feature selection strategy on top of leading VQA model features, we are able to extract 60 of the 763 statistical features used by the leading models.
Our experimental results show that VIDEVAL achieves state-of-theart performance at considerably lower computational cost than other leading models.
arXiv Detail & Related papers (2020-05-29T00:39:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.