HDR Video Reconstruction with a Large Dynamic Dataset in Raw and sRGB
Domains
- URL: http://arxiv.org/abs/2304.04773v2
- Date: Wed, 12 Apr 2023 04:06:16 GMT
- Title: HDR Video Reconstruction with a Large Dynamic Dataset in Raw and sRGB
Domains
- Authors: Huanjing Yue, Yubo Peng, Biting Yu, Xuanwu Yin, Zhenyu Zhou, Jingyu
Yang
- Abstract summary: High dynamic range ( HDR) video reconstruction is attracting more and more attention due to the superior visual quality compared with those of low dynamic range (LDR) videos.
There are still no real LDR- pairs for dynamic scenes due to the difficulty in capturing LDR- frames simultaneously.
In this work, we propose to utilize a staggered sensor to capture two alternate exposure images simultaneously, which are then fused into an HDR frame in both raw and sRGB domains.
- Score: 23.309488653045026
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: High dynamic range (HDR) video reconstruction is attracting more and more
attention due to the superior visual quality compared with those of low dynamic
range (LDR) videos. The availability of LDR-HDR training pairs is essential for
the HDR reconstruction quality. However, there are still no real LDR-HDR pairs
for dynamic scenes due to the difficulty in capturing LDR-HDR frames
simultaneously. In this work, we propose to utilize a staggered sensor to
capture two alternate exposure images simultaneously, which are then fused into
an HDR frame in both raw and sRGB domains. In this way, we build a large scale
LDR-HDR video dataset with 85 scenes and each scene contains 60 frames. Based
on this dataset, we further propose a Raw-HDRNet, which utilizes the raw LDR
frames as inputs. We propose a pyramid flow-guided deformation convolution to
align neighboring frames. Experimental results demonstrate that 1) the proposed
dataset can improve the HDR reconstruction performance on real scenes for three
benchmark networks; 2) Compared with sRGB inputs, utilizing raw inputs can
further improve the reconstruction quality and our proposed Raw-HDRNet is a
strong baseline for raw HDR reconstruction. Our dataset and code will be
released after the acceptance of this paper.
Related papers
- HDR-GS: Efficient High Dynamic Range Novel View Synthesis at 1000x Speed via Gaussian Splatting [76.5908492298286]
Existing HDR NVS methods are mainly based on NeRF.
They suffer from long training time and slow inference speed.
We propose a new framework, High Dynamic Range Gaussian Splatting (-GS)
arXiv Detail & Related papers (2024-05-24T00:46:58Z) - Self-Supervised High Dynamic Range Imaging with Multi-Exposure Images in
Dynamic Scenes [58.66427721308464]
Self is a self-supervised reconstruction method that only requires dynamic multi-exposure images during training.
Self achieves superior results against the state-of-the-art self-supervised methods, and comparable performance to supervised ones.
arXiv Detail & Related papers (2023-10-03T07:10:49Z) - Towards Efficient SDRTV-to-HDRTV by Learning from Image Formation [51.26219245226384]
Modern displays are capable of rendering video content with high dynamic range (WCG) and wide color gamut (SDR)
The majority of available resources are still in standard dynamic range (SDR)
We define and analyze the SDRTV-to-TV task by modeling the formation of SDRTV/TV content.
Our method is primarily designed for ultra-high-definition TV content and is therefore effective and lightweight for processing 4K resolution images.
arXiv Detail & Related papers (2023-09-08T02:50:54Z) - RawHDR: High Dynamic Range Image Reconstruction from a Single Raw Image [36.17182977927645]
High dynamic range (RGB) images capture much more intensity levels than standard ones.
Current methods predominantly generate HDR images from 8-bit low dynamic range (LDR) s images that have been degraded by the camera processing pipeline.
Unlike existing methods, the core idea of this work is to incorporate more informative Raw sensor data to generate HDR images.
arXiv Detail & Related papers (2023-09-05T07:58:21Z) - Multi-Bracket High Dynamic Range Imaging with Event Cameras [46.81570594990517]
We propose the first multi-bracket HDR pipeline combining a standard camera with an event camera.
Our results show better overall robustness when using events, with improvements in PSNR by up to 5dB on synthetic data and up to 0.7dB on real-world data.
arXiv Detail & Related papers (2022-03-13T11:10:47Z) - A Two-stage Deep Network for High Dynamic Range Image Reconstruction [0.883717274344425]
This study tackles the challenges of single-shot LDR to HDR mapping by proposing a novel two-stage deep network.
Notably, our proposed method aims to reconstruct an HDR image without knowing hardware information, including camera response function (CRF) and exposure settings.
arXiv Detail & Related papers (2021-04-19T15:19:17Z) - HDR Video Reconstruction: A Coarse-to-fine Network and A Real-world
Benchmark Dataset [30.249052175655606]
We introduce a coarse-to-fine deep learning framework for HDR video reconstruction.
Firstly, we perform coarse alignment and pixel blending in the image space to estimate the coarse HDR video.
Secondly, we conduct more sophisticated alignment and temporal fusion in the feature space of the coarse HDR video to produce better reconstruction.
arXiv Detail & Related papers (2021-03-27T16:40:05Z) - Beyond Visual Attractiveness: Physically Plausible Single Image HDR
Reconstruction for Spherical Panoramas [60.24132321381606]
We introduce the physical illuminance constraints to our single-shot HDR reconstruction framework.
Our method can generate HDRs which are not only visually appealing but also physically plausible.
arXiv Detail & Related papers (2021-03-24T01:51:19Z) - HDR-GAN: HDR Image Reconstruction from Multi-Exposed LDR Images with
Large Motions [62.44802076971331]
We propose a novel GAN-based model, HDR-GAN, for synthesizing HDR images from multi-exposed LDR images.
By incorporating adversarial learning, our method is able to produce faithful information in the regions with missing content.
arXiv Detail & Related papers (2020-07-03T11:42:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.