Unified Quality Assessment of In-the-Wild Videos with Mixed Datasets
Training
- URL: http://arxiv.org/abs/2011.04263v2
- Date: Sun, 15 Nov 2020 09:13:58 GMT
- Title: Unified Quality Assessment of In-the-Wild Videos with Mixed Datasets
Training
- Authors: Dingquan Li, Tingting Jiang, Ming Jiang
- Abstract summary: We focus on automatically assessing the quality of in-the-wild videos in computer vision applications.
To improve the performance of quality assessment models, we borrow intuitions from human perception.
We propose a mixed datasets training strategy for training a single VQA model with multiple datasets.
- Score: 20.288424566444224
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Video quality assessment (VQA) is an important problem in computer vision.
The videos in computer vision applications are usually captured in the wild. We
focus on automatically assessing the quality of in-the-wild videos, which is a
challenging problem due to the absence of reference videos, the complexity of
distortions, and the diversity of video contents. Moreover, the video contents
and distortions among existing datasets are quite different, which leads to
poor performance of data-driven methods in the cross-dataset evaluation
setting. To improve the performance of quality assessment models, we borrow
intuitions from human perception, specifically, content dependency and
temporal-memory effects of human visual system. To face the cross-dataset
evaluation challenge, we explore a mixed datasets training strategy for
training a single VQA model with multiple datasets. The proposed unified
framework explicitly includes three stages: relative quality assessor,
nonlinear mapping, and dataset-specific perceptual scale alignment, to jointly
predict relative quality, perceptual quality, and subjective quality.
Experiments are conducted on four publicly available datasets for VQA in the
wild, i.e., LIVE-VQC, LIVE-Qualcomm, KoNViD-1k, and CVD2014. The experimental
results verify the effectiveness of the mixed datasets training strategy and
prove the superior performance of the unified model in comparison with the
state-of-the-art models. For reproducible research, we make the PyTorch
implementation of our method available at https://github.com/lidq92/MDTVSFA.
Related papers
- Video Quality Assessment: A Comprehensive Survey [55.734935003021576]
Video quality assessment (VQA) is an important processing task, aiming at predicting the quality of videos in a manner consistent with human judgments of perceived quality.
We present a survey of recent progress in the development of VQA algorithms and the benchmarking studies and databases that make them possible.
arXiv Detail & Related papers (2024-12-04T05:25:17Z) - AIM 2024 Challenge on Compressed Video Quality Assessment: Methods and Results [120.95863275142727]
This paper presents the results of the Compressed Video Quality Assessment challenge, held in conjunction with the Advances in Image Manipulation (AIM) workshop at ECCV 2024.
The challenge aimed to evaluate the performance of VQA methods on a diverse dataset of 459 videos encoded with 14 codecs of various compression standards.
arXiv Detail & Related papers (2024-08-21T20:32:45Z) - Benchmarking Multi-dimensional AIGC Video Quality Assessment: A Dataset and Unified Model [56.03592388332793]
We investigate the AIGC-VQA problem, considering both subjective and objective quality assessment perspectives.
For the subjective perspective, we construct the Large-scale Generated Video Quality assessment (LGVQ) dataset, consisting of 2,808 AIGC videos.
We evaluate the perceptual quality of AIGC videos from three critical dimensions: spatial quality, temporal quality, and text-video alignment.
We propose the Unify Generated Video Quality assessment (UGVQ) model, designed to accurately evaluate the multi-dimensional quality of AIGC videos.
arXiv Detail & Related papers (2024-07-31T07:54:26Z) - CLIPVQA:Video Quality Assessment via CLIP [56.94085651315878]
We propose an efficient CLIP-based Transformer method for the VQA problem ( CLIPVQA)
The proposed CLIPVQA achieves new state-of-the-art VQA performance and up to 37% better generalizability than existing benchmark VQA methods.
arXiv Detail & Related papers (2024-07-06T02:32:28Z) - PTM-VQA: Efficient Video Quality Assessment Leveraging Diverse PreTrained Models from the Wild [27.195339506769457]
Video quality assessment (VQA) is a challenging problem due to the numerous factors that can affect the perceptual quality of a video.
Annotating the Mean opinion score (MOS) for videos is expensive and time-consuming, which limits the scale of VQA datasets.
We propose a VQA method named PTM-VQA, which leverages PreTrained Models to transfer knowledge from models pretrained on various pre-tasks.
arXiv Detail & Related papers (2024-05-28T02:37:29Z) - Analysis of Video Quality Datasets via Design of Minimalistic Video Quality Models [71.06007696593704]
Blind quality assessment (BVQA) plays an indispensable role in monitoring and improving the end-users' viewing experience in real-world video-enabled media applications.
As an experimental field, the improvements of BVQA models have been measured primarily on a few human-rated VQA datasets.
We conduct a first-of-its-kind computational analysis of VQA datasets via minimalistic BVQA models.
arXiv Detail & Related papers (2023-07-26T06:38:33Z) - Learning from Mixed Datasets: A Monotonic Image Quality Assessment Model [17.19991754976893]
We propose a monotonic neural network for IQA model learning with different datasets combined.
In particular, our model consists of a dataset-shared quality regressor and several dataset-specific quality transformers.
arXiv Detail & Related papers (2022-09-21T15:53:59Z) - CONVIQT: Contrastive Video Quality Estimator [63.749184706461826]
Perceptual video quality assessment (VQA) is an integral component of many streaming and video sharing platforms.
Here we consider the problem of learning perceptually relevant video quality representations in a self-supervised manner.
Our results indicate that compelling representations with perceptual bearing can be obtained using self-supervised learning.
arXiv Detail & Related papers (2022-06-29T15:22:01Z) - Blindly Assess Quality of In-the-Wild Videos via Quality-aware
Pre-training and Motion Perception [32.87570883484805]
We propose to transfer knowledge from image quality assessment (IQA) databases with authentic distortions and large-scale action recognition with rich motion patterns.
We train the proposed model on the target VQA databases using a mixed list-wise ranking loss function.
arXiv Detail & Related papers (2021-08-19T05:29:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.