Deep Reformulated Laplacian Tone Mapping
- URL: http://arxiv.org/abs/2102.00348v1
- Date: Sun, 31 Jan 2021 01:18:20 GMT
- Title: Deep Reformulated Laplacian Tone Mapping
- Authors: Jie Yang, Ziyi Liu, Mengchen Lin, Svetlana Yanushkevich, Orly
Yadid-Pecht
- Abstract summary: Wide dynamic range (WDR) images contain more scene details and contrast when compared to common images.
The details of WDR images can diminish during the tone mapping process.
In this work, we address the problem by combining a novel reformulated Laplacian pyramid and deep learning.
- Score: 6.078183247169192
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Wide dynamic range (WDR) images contain more scene details and contrast when
compared to common images. However, it requires tone mapping to process the
pixel values in order to display properly. The details of WDR images can
diminish during the tone mapping process. In this work, we address the problem
by combining a novel reformulated Laplacian pyramid and deep learning. The
reformulated Laplacian pyramid always decompose a WDR image into two frequency
bands where the low-frequency band is global feature-oriented, and the
high-frequency band is local feature-oriented. The reformulation preserves the
local features in its original resolution and condenses the global features
into a low-resolution image. The generated frequency bands are reconstructed
and fine-tuned to output the final tone mapped image that can display on the
screen with minimum detail and contrast loss. The experimental results
demonstrate that the proposed method outperforms state-of-the-art WDR image
tone mapping methods. The code is made publicly available at
https://github.com/linmc86/Deep-Reformulated-Laplacian-Tone-Mapping.
Related papers
- Pixel-Aligned Multi-View Generation with Depth Guided Decoder [86.1813201212539]
We propose a novel method for pixel-level image-to-multi-view generation.
Unlike prior work, we incorporate attention layers across multi-view images in the VAE decoder of a latent video diffusion model.
Our model enables better pixel alignment across multi-view images.
arXiv Detail & Related papers (2024-08-26T04:56:41Z) - Seeing the Unseen: A Frequency Prompt Guided Transformer for Image Restoration [44.35766203309201]
We develop a Frequency Prompting image restoration method, dubbed FPro, which can effectively provide prompt components from a frequency perspective to guild the restoration model.
Experimental results on commonly used benchmarks have demonstrated the favorable performance of our pipeline against SOTA methods on 5 image restoration tasks.
arXiv Detail & Related papers (2024-03-30T08:42:34Z) - Hi-Map: Hierarchical Factorized Radiance Field for High-Fidelity
Monocular Dense Mapping [51.739466714312805]
We introduce Hi-Map, a novel monocular dense mapping approach based on Neural Radiance Field (NeRF)
Hi-Map is exceptional in its capacity to achieve efficient and high-fidelity mapping using only posed RGB inputs.
arXiv Detail & Related papers (2024-01-06T12:32:25Z) - Lookup Table meets Local Laplacian Filter: Pyramid Reconstruction
Network for Tone Mapping [35.47139372780014]
This paper explores a novel strategy that integrates global and local operators by utilizing closed-form Laplacian pyramid decomposition and reconstruction.
We employ image-adaptive 3D LUTs to manipulate the tone in the low-frequency image by leveraging the specific characteristics of the frequency information.
We also utilize local Laplacian filters to refine the edge details in the high-frequency components in an adaptive manner.
arXiv Detail & Related papers (2023-10-26T07:05:38Z) - Symmetric Uncertainty-Aware Feature Transmission for Depth
Super-Resolution [52.582632746409665]
We propose a novel Symmetric Uncertainty-aware Feature Transmission (SUFT) for color-guided DSR.
Our method achieves superior performance compared to state-of-the-art methods.
arXiv Detail & Related papers (2023-06-01T06:35:59Z) - Unsupervised HDR Image and Video Tone Mapping via Contrastive Learning [19.346284003982035]
We propose a unified framework (IVTMNet) for unsupervised image and video tone mapping.
For video tone mapping, we propose a temporal-feature-replaced (TFR) module to efficiently utilize the temporal correlation.
Experimental results demonstrate that our method outperforms state-of-the-art image and video tone mapping methods.
arXiv Detail & Related papers (2023-03-13T17:45:39Z) - PINs: Progressive Implicit Networks for Multi-Scale Neural
Representations [68.73195473089324]
We propose a progressive positional encoding, exposing a hierarchical structure to incremental sets of frequency encodings.
Our model accurately reconstructs scenes with wide frequency bands and learns a scene representation at progressive level of detail.
Experiments on several 2D and 3D datasets show improvements in reconstruction accuracy, representational capacity and training speed compared to baselines.
arXiv Detail & Related papers (2022-02-09T20:33:37Z) - Wavelet-Based Network For High Dynamic Range Imaging [64.66969585951207]
Existing methods, such as optical flow based and end-to-end deep learning based solutions, are error-prone either in detail restoration or ghosting artifacts removal.
In this work, we propose a novel frequency-guided end-to-end deep neural network (FNet) to conduct HDR fusion in the frequency domain, and Wavelet Transform (DWT) is used to decompose inputs into different frequency bands.
The low-frequency signals are used to avoid specific ghosting artifacts, while the high-frequency signals are used for preserving details.
arXiv Detail & Related papers (2021-08-03T12:26:33Z) - WaveFill: A Wavelet-based Generation Network for Image Inpainting [57.012173791320855]
WaveFill is a wavelet-based inpainting network that decomposes images into multiple frequency bands.
WaveFill decomposes images by using discrete wavelet transform (DWT) that preserves spatial information naturally.
It applies L1 reconstruction loss to the low-frequency bands and adversarial loss to high-frequency bands, hence effectively mitigate inter-frequency conflicts.
arXiv Detail & Related papers (2021-07-23T04:44:40Z) - High-Resolution Photorealistic Image Translation in Real-Time: A
Laplacian Pyramid Translation Network [23.981019687483506]
We focus on speeding-up the high-resolution photorealistic I2IT tasks based on closed-form Laplacian pyramid decomposition and reconstruction.
We propose a Laplacian Pyramid Translation Network (N) to simultaneously perform these two tasks.
Our model avoids most of the heavy computation consumed by processing high-resolution feature maps and faithfully preserves the image details.
arXiv Detail & Related papers (2021-05-19T15:05:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.