A Lightweight Parallel Framework for Blind Image Quality Assessment
- URL: http://arxiv.org/abs/2402.12043v1
- Date: Mon, 19 Feb 2024 10:56:58 GMT
- Title: A Lightweight Parallel Framework for Blind Image Quality Assessment
- Authors: Qunyue Huang, Bin Fang
- Abstract summary: We propose a lightweight parallel framework (LPF) for blind image quality assessment (BIQA)
First, we extract the visual features using a pre-trained feature extraction network. Furthermore, we construct a simple yet effective feature embedding network (FEN) to transform the visual features.
We present two novel self-supervised subtasks, including a sample-level category prediction task and a batch-level quality comparison task.
- Score: 7.9562077122537875
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Existing blind image quality assessment (BIQA) methods focus on designing
complicated networks based on convolutional neural networks (CNNs) or
transformer. In addition, some BIQA methods enhance the performance of the
model in a two-stage training manner. Despite the significant advancements,
these methods remarkably raise the parameter count of the model, thus requiring
more training time and computational resources. To tackle the above issues, we
propose a lightweight parallel framework (LPF) for BIQA. First, we extract the
visual features using a pre-trained feature extraction network. Furthermore, we
construct a simple yet effective feature embedding network (FEN) to transform
the visual features, aiming to generate the latent representations that contain
salient distortion information. To improve the robustness of the latent
representations, we present two novel self-supervised subtasks, including a
sample-level category prediction task and a batch-level quality comparison
task. The sample-level category prediction task is presented to help the model
with coarse-grained distortion perception. The batch-level quality comparison
task is formulated to enhance the training data and thus improve the robustness
of the latent representations. Finally, the latent representations are fed into
a distortion-aware quality regression network (DaQRN), which simulates the
human vision system (HVS) and thus generates accurate quality scores.
Experimental results on multiple benchmark datasets demonstrate that the
proposed method achieves superior performance over state-of-the-art approaches.
Moreover, extensive analyses prove that the proposed method has lower
computational complexity and faster convergence speed.
Related papers
- Multi-Modal Prompt Learning on Blind Image Quality Assessment [65.0676908930946]
Image Quality Assessment (IQA) models benefit significantly from semantic information, which allows them to treat different types of objects distinctly.
Traditional methods, hindered by a lack of sufficiently annotated data, have employed the CLIP image-text pretraining model as their backbone to gain semantic awareness.
Recent approaches have attempted to address this mismatch using prompt technology, but these solutions have shortcomings.
This paper introduces an innovative multi-modal prompt-based methodology for IQA.
arXiv Detail & Related papers (2024-04-23T11:45:32Z) - GraFIQs: Face Image Quality Assessment Using Gradient Magnitudes [9.170455788675836]
Face Image Quality Assessment (FIQA) estimates the utility of face images for automated face recognition (FR) systems.
We propose in this work a novel approach to assess the quality of face images based on inspecting the required changes in the pre-trained FR model weights.
arXiv Detail & Related papers (2024-04-18T14:07:08Z) - Contrastive Pre-Training with Multi-View Fusion for No-Reference Point Cloud Quality Assessment [49.36799270585947]
No-reference point cloud quality assessment (NR-PCQA) aims to automatically evaluate the perceptual quality of distorted point clouds without available reference.
We propose a novel contrastive pre-training framework tailored for PCQA (CoPA)
Our method outperforms the state-of-the-art PCQA methods on popular benchmarks.
arXiv Detail & Related papers (2024-03-15T07:16:07Z) - Test Time Adaptation for Blind Image Quality Assessment [20.50795362928567]
We introduce two novel quality-relevant auxiliary tasks at the batch and sample levels to enable TTA for blind IQA.
Our experiments reveal that even using a small batch of images from the test distribution helps achieve significant improvement in performance.
arXiv Detail & Related papers (2023-07-27T09:43:06Z) - Learning Transformer Features for Image Quality Assessment [53.51379676690971]
We propose a unified IQA framework that utilizes CNN backbone and transformer encoder to extract features.
The proposed framework is compatible with both FR and NR modes and allows for a joint training scheme.
arXiv Detail & Related papers (2021-12-01T13:23:00Z) - A Practical Contrastive Learning Framework for Single-Image
Super-Resolution [51.422185656787285]
We investigate contrastive learning-based single image super-resolution from two perspectives.
We propose a practical contrastive learning framework for SISR, named PCL-SR.
Compared with existing benchmark methods, we re-train them by our proposed PCL-SR framework and achieve superior performance.
arXiv Detail & Related papers (2021-11-27T15:42:12Z) - Task-Specific Normalization for Continual Learning of Blind Image
Quality Models [105.03239956378465]
We present a simple yet effective continual learning method for blind image quality assessment (BIQA)
The key step in our approach is to freeze all convolution filters of a pre-trained deep neural network (DNN) for an explicit promise of stability.
We assign each new IQA dataset (i.e., task) a prediction head, and load the corresponding normalization parameters to produce a quality score.
The final quality estimate is computed by black a weighted summation of predictions from all heads with a lightweight $K$-means gating mechanism.
arXiv Detail & Related papers (2021-07-28T15:21:01Z) - Uncertainty-Aware Blind Image Quality Assessment in the Laboratory and
Wild [98.48284827503409]
We develop a textitunified BIQA model and an approach of training it for both synthetic and realistic distortions.
We employ the fidelity loss to optimize a deep neural network for BIQA over a large number of such image pairs.
Experiments on six IQA databases show the promise of the learned method in blindly assessing image quality in the laboratory and wild.
arXiv Detail & Related papers (2020-05-28T13:35:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.