HDR-VDP-3: A multi-metric for predicting image differences, quality and
contrast distortions in high dynamic range and regular content
- URL: http://arxiv.org/abs/2304.13625v1
- Date: Wed, 26 Apr 2023 15:32:04 GMT
- Title: HDR-VDP-3: A multi-metric for predicting image differences, quality and
contrast distortions in high dynamic range and regular content
- Authors: Rafal K. Mantiuk, Dounia Hammou, Param Hanji
- Abstract summary: High-Dynamic-Range Visual-Difference-Predictor version 3, or HDR-VDP-3, is a visual metric that can fulfill several tasks.
Here we present a high-level overview of the metric, position it with respect to related work, explain the main differences compared to version 2.2, and describe how the metric was adapted for the HDR Video Quality Measurement Grand Challenge 2023.
- Score: 14.75838951347139
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: High-Dynamic-Range Visual-Difference-Predictor version 3, or HDR-VDP-3, is a
visual metric that can fulfill several tasks, such as full-reference
image/video quality assessment, prediction of visual differences between a pair
of images, or prediction of contrast distortions. Here we present a high-level
overview of the metric, position it with respect to related work, explain the
main differences compared to version 2.2, and describe how the metric was
adapted for the HDR Video Quality Measurement Grand Challenge 2023.
Related papers
- No-Reference Point Cloud Quality Assessment via Graph Convolutional Network [89.12589881881082]
Three-dimensional (3D) point cloud, as an emerging visual media format, is increasingly favored by consumers.
Point clouds inevitably suffer from quality degradation and information loss through multimedia communication systems.
We propose a novel no-reference PCQA method by using a graph convolutional network (GCN) to characterize the mutual dependencies of multi-view 2D projected image contents.
arXiv Detail & Related papers (2024-11-12T11:39:05Z) - Dual-Branch Network for Portrait Image Quality Assessment [76.27716058987251]
We introduce a dual-branch network for portrait image quality assessment (PIQA)
We utilize two backbone networks (textiti.e., Swin Transformer-B) to extract the quality-aware features from the entire portrait image and the facial image cropped from it.
We leverage LIQE, an image scene classification and quality assessment model, to capture the quality-aware and scene-specific features as the auxiliary features.
arXiv Detail & Related papers (2024-05-14T12:43:43Z) - ColorVideoVDP: A visual difference predictor for image, video and display distortions [51.29162719944865]
metric is built on novel psychophysical models of chromatic contrast sensitivity and cross-channel contrast masking.
It accounts for the viewing conditions, geometric, and photometric characteristics of the display.
It was trained to predict common video streaming distortions and 8 new distortion types related to AR/VR displays.
arXiv Detail & Related papers (2024-01-21T13:16:33Z) - Perceptual Assessment and Optimization of HDR Image Rendering [25.72195917050074]
High dynamic range rendering has the ability to faithfully reproduce the wide luminance ranges in natural scenes.
Existing quality models are mostly designed for low dynamic range (LDR) images, and do not align well with human perception of HDR image quality.
We propose a family of HDR quality metrics, in which the key step is employing a simple inverse display model to decompose an HDR image into a stack of LDR images with varying exposures.
arXiv Detail & Related papers (2023-10-19T16:32:18Z) - Perceptual Tone Mapping Model for High Dynamic Range Imaging [0.0]
Tone mapping operators (TMOs) compress the luminance of HDR images without considering the surround and display conditions.
Current research addresses this challenge by incorporating perceptual color appearance attributes.
TMOz accounts for the effects of both the surround and the display conditions to achieve more optimal colorfulness reproduction.
arXiv Detail & Related papers (2023-09-29T04:45:48Z) - HDR or SDR? A Subjective and Objective Study of Scaled and Compressed
Videos [36.33823452846196]
We conducted a large-scale study of human perceptual quality judgments of High Dynamic Range (SDR) and Standard Dynamic Range (SDR) videos.
We found subject preference of HDR versus SDR depends heavily on the display device, as well as on resolution scaling and resolution.
arXiv Detail & Related papers (2023-04-25T21:43:37Z) - Deep Diversity-Enhanced Feature Representation of Hyperspectral Images [87.47202258194719]
We rectify 3D convolution by modifying its topology to enhance the rank upper-bound.
We also propose a novel diversity-aware regularization (DA-Reg) term that acts on the feature maps to maximize independence among elements.
To demonstrate the superiority of the proposed Re$3$-ConvSet and DA-Reg, we apply them to various HS image processing and analysis tasks.
arXiv Detail & Related papers (2023-01-15T16:19:18Z) - VIDM: Video Implicit Diffusion Models [75.90225524502759]
Diffusion models have emerged as a powerful generative method for synthesizing high-quality and diverse set of images.
We propose a video generation method based on diffusion models, where the effects of motion are modeled in an implicit condition.
We improve the quality of the generated videos by proposing multiple strategies such as sampling space truncation, robustness penalty, and positional group normalization.
arXiv Detail & Related papers (2022-12-01T02:58:46Z) - Locally Weighted Mean Phase Angle (LWMPA) Based Tone Mapping Quality
Index (TMQI-3) [0.0]
We propose a metric called the Tone Mapping Quality Index (TMQI-3), which evaluates the quality of the Low dynamic range (LDR) image based on its objective score.
TMQI-3 is noise resilient, takes account of structure and naturalness, and works on all three color channels combined into one component.
arXiv Detail & Related papers (2021-09-17T22:17:20Z) - Perceptual Quality Assessment of Omnidirectional Images as Moving Camera
Videos [49.217528156417906]
Two types of VR viewing conditions are crucial in determining the viewing behaviors of users and the perceived quality of the panorama.
We first transform an omnidirectional image to several video representations using different user viewing behaviors under different viewing conditions.
We then leverage advanced 2D full-reference video quality models to compute the perceived quality.
arXiv Detail & Related papers (2020-05-21T10:03:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.