Conformer and Blind Noisy Students for Improved Image Quality Assessment
- URL: http://arxiv.org/abs/2204.12819v1
- Date: Wed, 27 Apr 2022 10:21:08 GMT
- Title: Conformer and Blind Noisy Students for Improved Image Quality Assessment
- Authors: Marcos V. Conde, Maxime Burchi, Radu Timofte
- Abstract summary: Learning-based approaches for perceptual image quality assessment (IQA) usually require both the distorted and reference image for measuring the perceptual quality accurately.
In this work, we explore the performance of transformer-based full-reference IQA models.
We also propose a method for IQA based on semi-supervised knowledge distillation from full-reference teacher models into blind student models.
- Score: 80.57006406834466
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Generative models for image restoration, enhancement, and generation have
significantly improved the quality of the generated images. Surprisingly, these
models produce more pleasant images to the human eye than other methods, yet,
they may get a lower perceptual quality score using traditional perceptual
quality metrics such as PSNR or SSIM. Therefore, it is necessary to develop a
quantitative metric to reflect the performance of new algorithms, which should
be well-aligned with the person's mean opinion score (MOS). Learning-based
approaches for perceptual image quality assessment (IQA) usually require both
the distorted and reference image for measuring the perceptual quality
accurately. However, commonly only the distorted or generated image is
available. In this work, we explore the performance of transformer-based
full-reference IQA models. We also propose a method for IQA based on
semi-supervised knowledge distillation from full-reference teacher models into
blind student models using noisy pseudo-labeled data. Our approaches achieved
competitive results on the NTIRE 2022 Perceptual Image Quality Assessment
Challenge: our full-reference model was ranked 4th, and our blind noisy student
was ranked 3rd among 70 participants, each in their respective track.
Related papers
- Sliced Maximal Information Coefficient: A Training-Free Approach for Image Quality Assessment Enhancement [12.628718661568048]
We aim to explore a generalized human visual attention estimation strategy to mimic the process of human quality rating.
In particular, we model human attention generation by measuring the statistical dependency between the degraded image and the reference image.
Experimental results verify the performance of existing IQA models can be consistently improved when our attention module is incorporated.
arXiv Detail & Related papers (2024-08-19T11:55:32Z) - Multi-Modal Prompt Learning on Blind Image Quality Assessment [65.0676908930946]
Image Quality Assessment (IQA) models benefit significantly from semantic information, which allows them to treat different types of objects distinctly.
Traditional methods, hindered by a lack of sufficiently annotated data, have employed the CLIP image-text pretraining model as their backbone to gain semantic awareness.
Recent approaches have attempted to address this mismatch using prompt technology, but these solutions have shortcomings.
This paper introduces an innovative multi-modal prompt-based methodology for IQA.
arXiv Detail & Related papers (2024-04-23T11:45:32Z) - ARNIQA: Learning Distortion Manifold for Image Quality Assessment [28.773037051085318]
No-Reference Image Quality Assessment (NR-IQA) aims to develop methods to measure image quality in alignment with human perception without the need for a high-quality reference image.
We propose a self-supervised approach named ARNIQA for modeling the image distortion manifold to obtain quality representations in an intrinsic manner.
arXiv Detail & Related papers (2023-10-20T17:22:25Z) - Quality-aware Pre-trained Models for Blind Image Quality Assessment [15.566552014530938]
Blind image quality assessment (BIQA) aims to automatically evaluate the perceived quality of a single image.
In this paper, we propose to solve the problem by a pretext task customized for BIQA in a self-supervised learning manner.
arXiv Detail & Related papers (2023-03-01T13:52:40Z) - Perceptual Attacks of No-Reference Image Quality Models with
Human-in-the-Loop [113.75573175709573]
We make one of the first attempts to examine the perceptual robustness of NR-IQA models.
We test one knowledge-driven and three data-driven NR-IQA methods under four full-reference IQA models.
We find that all four NR-IQA models are vulnerable to the proposed perceptual attack.
arXiv Detail & Related papers (2022-10-03T13:47:16Z) - Confusing Image Quality Assessment: Towards Better Augmented Reality
Experience [96.29124666702566]
We consider AR technology as the superimposition of virtual scenes and real scenes, and introduce visual confusion as its basic theory.
A ConFusing Image Quality Assessment (CFIQA) database is established, which includes 600 reference images and 300 distorted images generated by mixing reference images in pairs.
An objective metric termed CFIQA is also proposed to better evaluate the confusing image quality.
arXiv Detail & Related papers (2022-04-11T07:03:06Z) - Image Quality Assessment using Contrastive Learning [50.265638572116984]
We train a deep Convolutional Neural Network (CNN) using a contrastive pairwise objective to solve the auxiliary problem.
We show through extensive experiments that CONTRIQUE achieves competitive performance when compared to state-of-the-art NR image quality models.
Our results suggest that powerful quality representations with perceptual relevance can be obtained without requiring large labeled subjective image quality datasets.
arXiv Detail & Related papers (2021-10-25T21:01:00Z) - No-Reference Image Quality Assessment via Feature Fusion and Multi-Task
Learning [29.19484863898778]
Blind or no-reference image quality assessment (NR-IQA) is a fundamental, unsolved, and yet challenging problem.
We propose a simple and yet effective general-purpose no-reference (NR) image quality assessment framework based on multi-task learning.
Our model employs distortion types as well as subjective human scores to predict image quality.
arXiv Detail & Related papers (2020-06-06T05:04:10Z) - Uncertainty-Aware Blind Image Quality Assessment in the Laboratory and
Wild [98.48284827503409]
We develop a textitunified BIQA model and an approach of training it for both synthetic and realistic distortions.
We employ the fidelity loss to optimize a deep neural network for BIQA over a large number of such image pairs.
Experiments on six IQA databases show the promise of the learned method in blindly assessing image quality in the laboratory and wild.
arXiv Detail & Related papers (2020-05-28T13:35:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.