Helping Visually Impaired People Take Better Quality Pictures
- URL: http://arxiv.org/abs/2305.08066v1
- Date: Sun, 14 May 2023 04:37:53 GMT
- Title: Helping Visually Impaired People Take Better Quality Pictures
- Authors: Maniratnam Mandal, Deepti Ghadiyaram, Danna Gurari, and Alan C. Bovik
- Abstract summary: We develop tools to help visually impaired users minimize occurrences of common technical distortions.
We also create a prototype feedback system that helps to guide users to mitigate quality issues.
- Score: 52.03016269364854
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Perception-based image analysis technologies can be used to help visually
impaired people take better quality pictures by providing automated guidance,
thereby empowering them to interact more confidently on social media. The
photographs taken by visually impaired users often suffer from one or both of
two kinds of quality issues: technical quality (distortions), and semantic
quality, such as framing and aesthetic composition. Here we develop tools to
help them minimize occurrences of common technical distortions, such as blur,
poor exposure, and noise. We do not address the complementary problems of
semantic quality, leaving that aspect for future work. The problem of assessing
and providing actionable feedback on the technical quality of pictures captured
by visually impaired users is hard enough, owing to the severe, commingled
distortions that often occur. To advance progress on the problem of analyzing
and measuring the technical quality of visually impaired user-generated content
(VI-UGC), we built a very large and unique subjective image quality and
distortion dataset. This new perceptual resource, which we call the LIVE-Meta
VI-UGC Database, contains $40$K real-world distorted VI-UGC images and $40$K
patches, on which we recorded $2.7$M human perceptual quality judgments and
$2.7$M distortion labels. Using this psychometric resource we also created an
automatic blind picture quality and distortion predictor that learns
local-to-global spatial quality relationships, achieving state-of-the-art
prediction performance on VI-UGC pictures, significantly outperforming existing
picture quality models on this unique class of distorted picture data. We also
created a prototype feedback system that helps to guide users to mitigate
quality issues and take better quality pictures, by creating a multi-task
learning framework.
Related papers
- Dual-Representation Interaction Driven Image Quality Assessment with Restoration Assistance [11.983231834400698]
No-Reference Image Quality Assessment for distorted images has always been a challenging problem due to image content variance and distortion diversity.
Previous IQA models mostly encode explicit single-quality features of synthetic images to obtain quality-aware representations for quality score prediction.
We introduce the DRI method to obtain degradation vectors and quality vectors of images, which separately model the degradation and quality information of low-quality images.
arXiv Detail & Related papers (2024-11-26T12:48:47Z) - Q-Ground: Image Quality Grounding with Large Multi-modality Models [61.72022069880346]
We introduce Q-Ground, the first framework aimed at tackling fine-scale visual quality grounding.
Q-Ground combines large multi-modality models with detailed visual quality analysis.
Central to our contribution is the introduction of the QGround-100K dataset.
arXiv Detail & Related papers (2024-07-24T06:42:46Z) - DSL-FIQA: Assessing Facial Image Quality via Dual-Set Degradation Learning and Landmark-Guided Transformer [23.70791030264281]
Generic Face Image Quality Assessment (GFIQA) evaluates the perceptual quality of facial images.
We present a novel transformer-based method for GFIQA, which is aided by two unique mechanisms.
arXiv Detail & Related papers (2024-06-13T23:11:25Z) - Dual-Branch Network for Portrait Image Quality Assessment [76.27716058987251]
We introduce a dual-branch network for portrait image quality assessment (PIQA)
We utilize two backbone networks (textiti.e., Swin Transformer-B) to extract the quality-aware features from the entire portrait image and the facial image cropped from it.
We leverage LIQE, an image scene classification and quality assessment model, to capture the quality-aware and scene-specific features as the auxiliary features.
arXiv Detail & Related papers (2024-05-14T12:43:43Z) - Quality-aware Pre-trained Models for Blind Image Quality Assessment [15.566552014530938]
Blind image quality assessment (BIQA) aims to automatically evaluate the perceived quality of a single image.
In this paper, we propose to solve the problem by a pretext task customized for BIQA in a self-supervised learning manner.
arXiv Detail & Related papers (2023-03-01T13:52:40Z) - MSTRIQ: No Reference Image Quality Assessment Based on Swin Transformer
with Multi-Stage Fusion [8.338999282303755]
We propose a novel algorithm based on the Swin Transformer.
It aggregates information from both local and global features to better predict the quality.
It ranks 2nd in the no-reference track of NTIRE 2022 Perceptual Image Quality Assessment Challenge.
arXiv Detail & Related papers (2022-05-20T11:34:35Z) - Conformer and Blind Noisy Students for Improved Image Quality Assessment [80.57006406834466]
Learning-based approaches for perceptual image quality assessment (IQA) usually require both the distorted and reference image for measuring the perceptual quality accurately.
In this work, we explore the performance of transformer-based full-reference IQA models.
We also propose a method for IQA based on semi-supervised knowledge distillation from full-reference teacher models into blind student models.
arXiv Detail & Related papers (2022-04-27T10:21:08Z) - Image Quality Assessment using Contrastive Learning [50.265638572116984]
We train a deep Convolutional Neural Network (CNN) using a contrastive pairwise objective to solve the auxiliary problem.
We show through extensive experiments that CONTRIQUE achieves competitive performance when compared to state-of-the-art NR image quality models.
Our results suggest that powerful quality representations with perceptual relevance can be obtained without requiring large labeled subjective image quality datasets.
arXiv Detail & Related papers (2021-10-25T21:01:00Z) - Towards Unsupervised Deep Image Enhancement with Generative Adversarial
Network [92.01145655155374]
We present an unsupervised image enhancement generative network (UEGAN)
It learns the corresponding image-to-image mapping from a set of images with desired characteristics in an unsupervised manner.
Results show that the proposed model effectively improves the aesthetic quality of images.
arXiv Detail & Related papers (2020-12-30T03:22:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.