PMT-IQA: Progressive Multi-task Learning for Blind Image Quality
Assessment
- URL: http://arxiv.org/abs/2301.01182v2
- Date: Fri, 3 Nov 2023 14:04:53 GMT
- Title: PMT-IQA: Progressive Multi-task Learning for Blind Image Quality
Assessment
- Authors: Qingyi Pan, Ning Guo, Letu Qingge, Jingyi Zhang, Pei Yang
- Abstract summary: Blind image quality assessment (BIQA) remains challenging due to the diversity of distortion and image content variation.
Existing BIQA methods often fail to consider multi-scale distortion patterns and image content.
We propose a Progressive Multi-Task Image Quality Assessment (PMT-IQA) model, which contains a multi-scale feature extraction module (MS) and a progressive multi-task learning module (PMT)
- Score: 6.976106651278988
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Blind image quality assessment (BIQA) remains challenging due to the
diversity of distortion and image content variation, which complicate the
distortion patterns crossing different scales and aggravate the difficulty of
the regression problem for BIQA. However, existing BIQA methods often fail to
consider multi-scale distortion patterns and image content, and little research
has been done on learning strategies to make the regression model produce
better performance. In this paper, we propose a simple yet effective
Progressive Multi-Task Image Quality Assessment (PMT-IQA) model, which contains
a multi-scale feature extraction module (MS) and a progressive multi-task
learning module (PMT), to help the model learn complex distortion patterns and
better optimize the regression issue to align with the law of human learning
process from easy to hard. To verify the effectiveness of the proposed PMT-IQA
model, we conduct experiments on four widely used public datasets, and the
experimental results indicate that the performance of PMT-IQA is superior to
the comparison approaches, and both MS and PMT modules improve the model's
performance.
Related papers
- MMEvol: Empowering Multimodal Large Language Models with Evol-Instruct [148.39859547619156]
We propose MMEvol, a novel multimodal instruction data evolution framework.
MMEvol iteratively improves data quality through a refined combination of fine-grained perception, cognitive reasoning, and interaction evolution.
Our approach reaches state-of-the-art (SOTA) performance in nine tasks using significantly less data compared to state-of-the-art models.
arXiv Detail & Related papers (2024-09-09T17:44:00Z) - MSLIQA: Enhancing Learning Representations for Image Quality Assessment through Multi-Scale Learning [6.074775040047959]
We improve the performance of a generic lightweight NR-IQA model by introducing a novel augmentation strategy.
This augmentation strategy enables the network to better discriminate between different distortions in various parts of the image by zooming in and out.
The inclusion of test-time augmentation further enhances performance, making our lightweight network's results comparable to the current state-of-the-art models.
arXiv Detail & Related papers (2024-08-29T20:05:02Z) - Opinion-Unaware Blind Image Quality Assessment using Multi-Scale Deep Feature Statistics [54.08757792080732]
We propose integrating deep features from pre-trained visual models with a statistical analysis model to achieve opinion-unaware BIQA (OU-BIQA)
Our proposed model exhibits superior consistency with human visual perception compared to state-of-the-art BIQA models.
arXiv Detail & Related papers (2024-05-29T06:09:34Z) - Contrastive Pre-Training with Multi-View Fusion for No-Reference Point Cloud Quality Assessment [49.36799270585947]
No-reference point cloud quality assessment (NR-PCQA) aims to automatically evaluate the perceptual quality of distorted point clouds without available reference.
We propose a novel contrastive pre-training framework tailored for PCQA (CoPA)
Our method outperforms the state-of-the-art PCQA methods on popular benchmarks.
arXiv Detail & Related papers (2024-03-15T07:16:07Z) - Comparison of No-Reference Image Quality Models via MAP Estimation in
Diffusion Latents [99.19391983670569]
We show that NR-IQA models can be plugged into the maximum a posteriori (MAP) estimation framework for image enhancement.
Different NR-IQA models are likely to induce different enhanced images, which are ultimately subject to psychophysical testing.
This leads to a new computational method for comparing NR-IQA models within the analysis-by-synthesis framework.
arXiv Detail & Related papers (2024-03-11T03:35:41Z) - Diffusion Model Based Visual Compensation Guidance and Visual Difference
Analysis for No-Reference Image Quality Assessment [82.13830107682232]
We propose a novel class of state-of-the-art (SOTA) generative model, which exhibits the capability to model intricate relationships.
We devise a new diffusion restoration network that leverages the produced enhanced image and noise-containing images.
Two visual evaluation branches are designed to comprehensively analyze the obtained high-level feature information.
arXiv Detail & Related papers (2024-02-22T09:39:46Z) - A Lightweight Parallel Framework for Blind Image Quality Assessment [7.9562077122537875]
We propose a lightweight parallel framework (LPF) for blind image quality assessment (BIQA)
First, we extract the visual features using a pre-trained feature extraction network. Furthermore, we construct a simple yet effective feature embedding network (FEN) to transform the visual features.
We present two novel self-supervised subtasks, including a sample-level category prediction task and a batch-level quality comparison task.
arXiv Detail & Related papers (2024-02-19T10:56:58Z) - MD-IQA: Learning Multi-scale Distributed Image Quality Assessment with
Semi Supervised Learning for Low Dose CT [6.158876574189994]
Image quality assessment (IQA) plays a critical role in optimizing radiation dose and developing novel medical imaging techniques.
Recent deep learning-based approaches have demonstrated strong modeling capabilities and potential for medical IQA.
We propose a multi-scale distributions regression approach to predict quality scores by constraining the output distribution.
arXiv Detail & Related papers (2023-11-14T09:33:33Z) - No-Reference Image Quality Assessment via Feature Fusion and Multi-Task
Learning [29.19484863898778]
Blind or no-reference image quality assessment (NR-IQA) is a fundamental, unsolved, and yet challenging problem.
We propose a simple and yet effective general-purpose no-reference (NR) image quality assessment framework based on multi-task learning.
Our model employs distortion types as well as subjective human scores to predict image quality.
arXiv Detail & Related papers (2020-06-06T05:04:10Z) - Uncertainty-Aware Blind Image Quality Assessment in the Laboratory and
Wild [98.48284827503409]
We develop a textitunified BIQA model and an approach of training it for both synthetic and realistic distortions.
We employ the fidelity loss to optimize a deep neural network for BIQA over a large number of such image pairs.
Experiments on six IQA databases show the promise of the learned method in blindly assessing image quality in the laboratory and wild.
arXiv Detail & Related papers (2020-05-28T13:35:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.