X2HDR: HDR Image Generation in a Perceptually Uniform Space
- URL: http://arxiv.org/abs/2602.04814v1
- Date: Wed, 04 Feb 2026 17:59:51 GMT
- Title: X2HDR: HDR Image Generation in a Perceptually Uniform Space
- Authors: Ronghuan Wu, Wanchao Su, Kede Ma, Jing Liao, RafaĆ K. Mantiuk,
- Abstract summary: High-dynamic-range formats and displays are becoming increasingly prevalent, yet state-of-the-art image generators are limited to low-dynamic-range (LDR) output.<n>We show that existing pretrained diffusion models can be easily adapted to HDR generation without retraining from scratch.
- Score: 37.83280929526874
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: High-dynamic-range (HDR) formats and displays are becoming increasingly prevalent, yet state-of-the-art image generators (e.g., Stable Diffusion and FLUX) typically remain limited to low-dynamic-range (LDR) output due to the lack of large-scale HDR training data. In this work, we show that existing pretrained diffusion models can be easily adapted to HDR generation without retraining from scratch. A key challenge is that HDR images are natively represented in linear RGB, whose intensity and color statistics differ substantially from those of sRGB-encoded LDR images. This gap, however, can be effectively bridged by converting HDR inputs into perceptually uniform encodings (e.g., using PU21 or PQ). Empirically, we find that LDR-pretrained variational autoencoders (VAEs) reconstruct PU21-encoded HDR inputs with fidelity comparable to LDR data, whereas linear RGB inputs cause severe degradations. Motivated by this finding, we describe an efficient adaptation strategy that freezes the VAE and finetunes only the denoiser via low-rank adaptation in a perceptually uniform space. This results in a unified computational method that supports both text-to-HDR synthesis and single-image RAW-to-HDR reconstruction. Experiments demonstrate that our perceptually encoded adaptation consistently improves perceptual fidelity, text-image alignment, and effective dynamic range, relative to previous techniques.
Related papers
- Learned HDR Image Compression for Perceptually Optimal Storage and Display [21.772946547671122]
High dynamic range capture and display have seen significant growth in popularity driven by the advancements in technology and increasing consumer demand for superior image quality.
As a result, HDR image compression is crucial to fully realize the benefits of HDR imaging without suffering from large file sizes and inefficient data handling.
In this work, we initiate efforts towards end-to-end optimized HDR image compression for perceptually optimal storage and display.
arXiv Detail & Related papers (2024-07-18T05:35:57Z) - HDR-GS: Efficient High Dynamic Range Novel View Synthesis at 1000x Speed via Gaussian Splatting [76.5908492298286]
Existing HDR NVS methods are mainly based on NeRF.
They suffer from long training time and slow inference speed.
We propose a new framework, High Dynamic Range Gaussian Splatting (-GS)
arXiv Detail & Related papers (2024-05-24T00:46:58Z) - Bracket Diffusion: HDR Image Generation by Consistent LDR Denoising [29.45922922270381]
We demonstrate generating HDR images using the concerted action of multiple black-box, pre-trained LDR image diffusion models.<n>We operate multiple denoising processes to generate multiple LDR brackets that together form a valid HDR result.<n>We demonstrate state-of-the-art unconditional and conditional restoration-type (LDR2) generative modeling results, yet in HDR.
arXiv Detail & Related papers (2024-05-23T08:24:22Z) - Adapting Pretrained Networks for Image Quality Assessment on High Dynamic Range Displays [0.0]
Conventional image quality metrics (IQMs) are designed for perceptually uniform gamma-encoded pixel values.
Most of the available datasets consist of standard-dynamic-range (SDR) images collected in standard and possibly uncontrolled viewing conditions.
Popular pre-trained neural networks are likewise intended for SDR inputs, restricting their direct application to HDR content.
In this work, we explore more effective approaches for training deep learning-based models for image quality assessment (IQA) on HDR data.
arXiv Detail & Related papers (2024-05-01T17:57:12Z) - Generating Content for HDR Deghosting from Frequency View [56.103761824603644]
Recent Diffusion Models (DMs) have been introduced in HDR imaging field.
DMs require extensive iterations with large models to estimate entire images.
We propose the Low-Frequency aware Diffusion (LF-Diff) model for ghost-free HDR imaging.
arXiv Detail & Related papers (2024-04-01T01:32:11Z) - HistoHDR-Net: Histogram Equalization for Single LDR to HDR Image
Translation [12.45632443397018]
High Dynamic Range ( HDR) imaging aims to replicate the high visual quality and clarity of real-world scenes.
The literature offers various data-driven methods for HDR image reconstruction from Low Dynamic Range (LDR) counterparts.
A common limitation of these approaches is missing details in regions of the reconstructed HDR images.
We propose a simple and effective method, Histo-Net, to recover the fine details.
arXiv Detail & Related papers (2024-02-08T20:14:46Z) - Training Neural Networks on RAW and HDR Images for Restoration Tasks [53.84872583527721]
We study how neural networks should be trained for tasks on RAW and HDR images in linear color spaces.<n>Our results indicate that neural networks train significantly better on HDR and RAW images represented in color spaces.<n>This small change to the training strategy can bring a very substantial gain in performance, between 2 and 9 dB.
arXiv Detail & Related papers (2023-12-06T17:47:16Z) - Towards High-quality HDR Deghosting with Conditional Diffusion Models [88.83729417524823]
High Dynamic Range (LDR) images can be recovered from several Low Dynamic Range (LDR) images by existing Deep Neural Networks (DNNs) techniques.
DNNs still generate ghosting artifacts when LDR images have saturation and large motion.
We formulate the HDR deghosting problem as an image generation that leverages LDR features as the diffusion model's condition.
arXiv Detail & Related papers (2023-11-02T01:53:55Z) - Efficient HDR Reconstruction from Real-World Raw Images [16.54071503000866]
High-definition screens on edge devices stimulate a strong demand for efficient high dynamic range ( HDR) algorithms.
Many existing HDR methods either deliver unsatisfactory results or consume too much computational and memory resources.
In this work, we discover an excellent opportunity for HDR reconstructing directly from raw images and investigating novel neural network structures.
arXiv Detail & Related papers (2023-06-17T10:10:15Z) - Invertible Tone Mapping with Selectable Styles [19.03179521805971]
In this paper, we propose an invertible tone mapping method that converts the multi-exposure HDR to a true LDR.
Our invertible LDR can mimic the appearance of a user-selected tone mapping style.
It can be shared over any existing social network platforms that may re-encode or format-convert the uploaded images.
arXiv Detail & Related papers (2021-10-09T07:32:36Z) - HDR-GAN: HDR Image Reconstruction from Multi-Exposed LDR Images with
Large Motions [62.44802076971331]
We propose a novel GAN-based model, HDR-GAN, for synthesizing HDR images from multi-exposed LDR images.
By incorporating adversarial learning, our method is able to produce faithful information in the regions with missing content.
arXiv Detail & Related papers (2020-07-03T11:42:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.