Training Neural Networks on RAW and HDR Images for Restoration Tasks
- URL: http://arxiv.org/abs/2312.03640v3
- Date: Sat, 19 Apr 2025 16:34:50 GMT
- Title: Training Neural Networks on RAW and HDR Images for Restoration Tasks
- Authors: Andrew Yanzhe Ke, Lei Luo, Xiaoyu Xiang, Yuchen Fan, Rakesh Ranjan, Alexandre Chapiro, RafaĆ K. Mantiuk,
- Abstract summary: We study how neural networks should be trained for tasks on RAW and HDR images in linear color spaces.<n>Our results indicate that neural networks train significantly better on HDR and RAW images represented in color spaces.<n>This small change to the training strategy can bring a very substantial gain in performance, between 2 and 9 dB.
- Score: 53.84872583527721
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The vast majority of standard image and video content available online is represented in display-encoded color spaces, in which pixel values are conveniently scaled to a limited range (0-1) and the color distribution is approximately perceptually uniform. In contrast, both camera RAW and high dynamic range (HDR) images are often represented in linear color spaces, in which color values are linearly related to colorimetric quantities of light. While training on commonly available display-encoded images is a well-established practice, there is no consensus on how neural networks should be trained for tasks on RAW and HDR images in linear color spaces. In this work, we test several approaches on three popular image restoration applications: denoising, deblurring, and single-image super-resolution. We examine whether HDR/RAW images need to be display-encoded using popular transfer functions (PQ, PU21, and mu-law), or whether it is better to train in linear color spaces, but use loss functions that correct for perceptual non-uniformity. Our results indicate that neural networks train significantly better on HDR and RAW images represented in display-encoded color spaces, which offer better perceptual uniformity than linear spaces. This small change to the training strategy can bring a very substantial gain in performance, between 2 and 9 dB.
Related papers
- X2HDR: HDR Image Generation in a Perceptually Uniform Space [37.83280929526874]
High-dynamic-range formats and displays are becoming increasingly prevalent, yet state-of-the-art image generators are limited to low-dynamic-range (LDR) output.<n>We show that existing pretrained diffusion models can be easily adapted to HDR generation without retraining from scratch.
arXiv Detail & Related papers (2026-02-04T17:59:51Z) - Log NeRF: Comparing Spaces for Learning Radiance Fields [0.6542188603141654]
Neural Radiance Fields (NeRF) have achieved remarkable results in novel view synthesis.<n>Inspired by the BiIlluminant Dichromatic Reflection (BIDR) model, we hypothesize that log RGB space enables NeRF to learn a more compact and effective representation of scene appearance.<n>We trained NeRF models under various color space interpretations, converting each network output to a common color space before rendering and loss computation, enforcing representation learning in different color spaces.
arXiv Detail & Related papers (2025-12-10T07:12:33Z) - You Only Need One Color Space: An Efficient Network for Low-light Image Enhancement [50.37253008333166]
Low-Light Image Enhancement (LLIE) task tends to restore the details and visual information from corrupted low-light images.
We propose a novel trainable color space, named Horizontal/Vertical-Intensity (HVI)
It not only decouples brightness and color from RGB channels to mitigate the instability during enhancement but also adapts to low-light images in different illumination ranges due to the trainable parameters.
arXiv Detail & Related papers (2024-02-08T16:47:43Z) - Self-Supervised High Dynamic Range Imaging with Multi-Exposure Images in
Dynamic Scenes [58.66427721308464]
Self is a self-supervised reconstruction method that only requires dynamic multi-exposure images during training.
Self achieves superior results against the state-of-the-art self-supervised methods, and comparable performance to supervised ones.
arXiv Detail & Related papers (2023-10-03T07:10:49Z) - Transform your Smartphone into a DSLR Camera: Learning the ISP in the
Wild [159.71025525493354]
We propose a trainable Image Signal Processing framework that produces DSLR quality images given RAW images captured by a smartphone.
To address the color misalignments between training image pairs, we employ a color-conditional ISP network and optimize a novel parametric color mapping between each input RAW and reference DSLR image.
arXiv Detail & Related papers (2022-03-20T20:13:59Z) - Astronomical Image Colorization and upscaling with Generative
Adversarial Networks [0.0]
This research aims to provide an automated approach for the problem by focusing on a very specific domain of images, namely astronomical images.
We explore the usage of various models in two different color spaces, RGB and L*a*b.
The model produces visually appealing images which hallucinate high resolution, colorized data in these results which does not exist in the original image.
arXiv Detail & Related papers (2021-12-27T19:01:20Z) - Locally Weighted Mean Phase Angle (LWMPA) Based Tone Mapping Quality
Index (TMQI-3) [0.0]
We propose a metric called the Tone Mapping Quality Index (TMQI-3), which evaluates the quality of the Low dynamic range (LDR) image based on its objective score.
TMQI-3 is noise resilient, takes account of structure and naturalness, and works on all three color channels combined into one component.
arXiv Detail & Related papers (2021-09-17T22:17:20Z) - Learning RAW-to-sRGB Mappings with Inaccurately Aligned Supervision [76.41657124981549]
This paper presents a joint learning model for image alignment and RAW-to-sRGB mapping.
Experiments show that our method performs favorably against state-of-the-arts on ZRR and SR-RAW datasets.
arXiv Detail & Related papers (2021-08-18T12:41:36Z) - Learning to Structure an Image with Few Colors [59.34619548026885]
We propose a color quantization network, ColorCNN, which learns to structure the images from the classification loss in an end-to-end manner.
With only a 1-bit color space (i.e., two colors), the proposed network achieves 82.1% top-1 accuracy on the CIFAR10 dataset.
For applications, when encoded with PNG, the proposed color quantization shows superiority over other image compression methods in the extremely low bit-rate regime.
arXiv Detail & Related papers (2020-03-17T17:56:15Z) - Adaptive Dithering Using Curved Markov-Gaussian Noise in the Quantized
Domain for Mapping SDR to HDR Image [2.913398015606848]
High Dynamic Range (SDR) imaging is gaining increased attention due to its realistic content, for not only regular displays but also smartphones.
We present a technique for noise generation that operates on pixels of a quantized image.
We vary the magnitude and structure of the noise pattern adaptively based on the luma of the quantized pixel and the slope of the inverse-tone mapping function.
arXiv Detail & Related papers (2020-01-20T05:30:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.