Deep Shape-Texture Statistics for Completely Blind Image Quality
Evaluation
- URL: http://arxiv.org/abs/2401.08107v1
- Date: Tue, 16 Jan 2024 04:28:09 GMT
- Title: Deep Shape-Texture Statistics for Completely Blind Image Quality
Evaluation
- Authors: Yixuan Li, Peilin Chen, Hanwei Zhu, Keyan Ding, Leida Li, and Shiqi
Wang
- Abstract summary: Deep features as visual descriptors have advanced IQA in recent research, but they are discovered to be highly texture-biased and lack of shape-bias.
We find out that image shape and texture cues respond differently towards distortions, and the absence of either one results in an incomplete image representation.
To formulate a well-round statistical description for images, we utilize the shapebiased and texture-biased deep features produced by Deep Neural Networks (DNNs) simultaneously.
- Score: 48.278380421089764
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Opinion-Unaware Blind Image Quality Assessment (OU-BIQA) models aim to
predict image quality without training on reference images and subjective
quality scores. Thereinto, image statistical comparison is a classic paradigm,
while the performance is limited by the representation ability of visual
descriptors. Deep features as visual descriptors have advanced IQA in recent
research, but they are discovered to be highly texture-biased and lack of
shape-bias. On this basis, we find out that image shape and texture cues
respond differently towards distortions, and the absence of either one results
in an incomplete image representation. Therefore, to formulate a well-round
statistical description for images, we utilize the shapebiased and
texture-biased deep features produced by Deep Neural Networks (DNNs)
simultaneously. More specifically, we design a Shape-Texture Adaptive Fusion
(STAF) module to merge shape and texture information, based on which we
formulate qualityrelevant image statistics. The perceptual quality is
quantified by the variant Mahalanobis Distance between the inner and outer
Shape-Texture Statistics (DSTS), wherein the inner and outer statistics
respectively describe the quality fingerprints of the distorted image and
natural images. The proposed DSTS delicately utilizes shape-texture statistical
relations between different data scales in the deep domain, and achieves
state-of-the-art (SOTA) quality prediction performance on images with
artificial and authentic distortions.
Related papers
- On quantifying and improving realism of images generated with diffusion [50.37578424163951]
We propose a metric, called Image Realism Score (IRS), computed from five statistical measures of a given image.
IRS is easily usable as a measure to classify a given image as real or fake.
We experimentally establish the model- and data-agnostic nature of the proposed IRS by successfully detecting fake images generated by Stable Diffusion Model (SDM), Dalle2, Midjourney and BigGAN.
Our efforts have also led to Gen-100 dataset, which provides 1,000 samples for 100 classes generated by four high-quality models.
arXiv Detail & Related papers (2023-09-26T08:32:55Z) - Training-free Diffusion Model Adaptation for Variable-Sized
Text-to-Image Synthesis [45.19847146506007]
Diffusion models (DMs) have recently gained attention with state-of-the-art performance in text-to-image synthesis.
This paper focuses on adapting text-to-image diffusion models to handle variety while maintaining visual fidelity.
arXiv Detail & Related papers (2023-06-14T17:23:07Z) - DeepDC: Deep Distance Correlation as a Perceptual Image Quality
Evaluator [53.57431705309919]
ImageNet pre-trained deep neural networks (DNNs) show notable transferability for building effective image quality assessment (IQA) models.
We develop a novel full-reference IQA (FR-IQA) model based exclusively on pre-trained DNN features.
We conduct comprehensive experiments to demonstrate the superiority of the proposed quality model on five standard IQA datasets.
arXiv Detail & Related papers (2022-11-09T14:57:27Z) - DeepWSD: Projecting Degradations in Perceptual Space to Wasserstein
Distance in Deep Feature Space [67.07476542850566]
We propose to model the quality degradation in perceptual space from a statistical distribution perspective.
The quality is measured based upon the Wasserstein distance in the deep feature domain.
The deep Wasserstein distance (DeepWSD) performed on features from neural networks enjoys better interpretability of the quality contamination.
arXiv Detail & Related papers (2022-08-05T02:46:12Z) - Image Quality Assessment using Contrastive Learning [50.265638572116984]
We train a deep Convolutional Neural Network (CNN) using a contrastive pairwise objective to solve the auxiliary problem.
We show through extensive experiments that CONTRIQUE achieves competitive performance when compared to state-of-the-art NR image quality models.
Our results suggest that powerful quality representations with perceptual relevance can be obtained without requiring large labeled subjective image quality datasets.
arXiv Detail & Related papers (2021-10-25T21:01:00Z) - Locally Adaptive Structure and Texture Similarity for Image Quality
Assessment [33.58928017067797]
We describe a locally adaptive structure and texture similarity index for full-reference image quality assessment (IQA)
Specifically, we rely on a single statistical feature, namely the dispersion index, to localize texture regions at different scales.
The resulting A-DISTS is adapted to local image content, and is free of expensive human perceptual scores for supervised training.
arXiv Detail & Related papers (2021-10-16T09:19:56Z) - A Deep Drift-Diffusion Model for Image Aesthetic Score Distribution
Prediction [68.76594695163386]
We propose a Deep Drift-Diffusion model inspired by psychologists to predict aesthetic score distribution from images.
The DDD model can describe the psychological process of aesthetic perception instead of traditional modeling of the results of assessment.
Our novel DDD model is simple but efficient, which outperforms the state-of-the-art methods in aesthetic score distribution prediction.
arXiv Detail & Related papers (2020-10-15T11:01:46Z) - Image Quality Assessment: Unifying Structure and Texture Similarity [38.05659069533254]
We develop the first full-reference image quality model with explicit tolerance to texture resampling.
Using a convolutional neural network, we construct an injective and differentiable function that transforms images to overcomplete representations.
arXiv Detail & Related papers (2020-04-16T16:11:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.