High Dynamic Range Imaging with Context-aware Transformer
- URL: http://arxiv.org/abs/2304.04416v4
- Date: Fri, 21 Apr 2023 06:56:13 GMT
- Title: High Dynamic Range Imaging with Context-aware Transformer
- Authors: Fangfang Zhou, Dan Zhang and Zhenming Fu
- Abstract summary: We propose a novel hierarchical dual Transformer (HDT) method for ghost-free images.
First, we use a CNN-based head with spatial attention mechanisms to extract features from all the LDR images.
Second, the LDR features are delivered to the Transformer, while the local details are extracted using the channel attention mechanism.
- Score: 3.1892103878735454
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Avoiding the introduction of ghosts when synthesising LDR images as high
dynamic range (HDR) images is a challenging task. Convolutional neural networks
(CNNs) are effective for HDR ghost removal in general, but are challenging to
deal with the LDR images if there are large movements or
oversaturation/undersaturation. Existing dual-branch methods combining CNN and
Transformer omit part of the information from non-reference images, while the
features extracted by the CNN-based branch are bound to the kernel size with
small receptive field, which are detrimental to the deblurring and the recovery
of oversaturated/undersaturated regions. In this paper, we propose a novel
hierarchical dual Transformer method for ghost-free HDR (HDT-HDR) images
generation, which extracts global features and local features simultaneously.
First, we use a CNN-based head with spatial attention mechanisms to extract
features from all the LDR images. Second, the LDR features are delivered to the
Hierarchical Dual Transformer (HDT). In each Dual Transformer (DT), the global
features are extracted by the window-based Transformer, while the local details
are extracted using the channel attention mechanism with deformable CNNs.
Finally, the ghost free HDR image is obtained by dimensional mapping on the HDT
output. Abundant experiments demonstrate that our HDT-HDR achieves the
state-of-the-art performance among existing HDR ghost removal methods.
Related papers
- FastHDRNet: A new efficient method for SDR-to-HDR Translation [5.224011800476952]
We propose a neural network for SDR to HDR conversion, termed "FastNet"
The architecture is designed as a lightweight network that utilizes global statistics and local information with super high efficiency.
arXiv Detail & Related papers (2024-04-06T03:25:24Z) - Towards High-quality HDR Deghosting with Conditional Diffusion Models [88.83729417524823]
High Dynamic Range (LDR) images can be recovered from several Low Dynamic Range (LDR) images by existing Deep Neural Networks (DNNs) techniques.
DNNs still generate ghosting artifacts when LDR images have saturation and large motion.
We formulate the HDR deghosting problem as an image generation that leverages LDR features as the diffusion model's condition.
arXiv Detail & Related papers (2023-11-02T01:53:55Z) - Self-Supervised High Dynamic Range Imaging with Multi-Exposure Images in
Dynamic Scenes [58.66427721308464]
Self is a self-supervised reconstruction method that only requires dynamic multi-exposure images during training.
Self achieves superior results against the state-of-the-art self-supervised methods, and comparable performance to supervised ones.
arXiv Detail & Related papers (2023-10-03T07:10:49Z) - A Unified HDR Imaging Method with Pixel and Patch Level [41.14378863436963]
We propose a hybrid HDR deghosting network, called HyNet, to generate visually pleasing HDR images.
Experiments demonstrate that HyNet outperforms state-of-the-art methods both quantitatively and qualitatively, achieving appealing HDR visualization with unified textures and colors.
arXiv Detail & Related papers (2023-04-14T06:21:57Z) - SMAE: Few-shot Learning for HDR Deghosting with Saturation-Aware Masked
Autoencoders [97.64072440883392]
We propose a novel semi-supervised approach to realize few-shot HDR imaging via two stages of training, called SSHDR.
Unlikely previous methods, directly recovering content and removing ghosts simultaneously, which is hard to achieve optimum.
Experiments demonstrate that SSHDR outperforms state-of-the-art methods quantitatively and qualitatively within and across different datasets.
arXiv Detail & Related papers (2023-04-14T03:42:51Z) - Ghost-free High Dynamic Range Imaging via Hybrid CNN-Transformer and
Structure Tensor [12.167049432063132]
We present a hybrid model consisting of a convolutional encoder and a Transformer decoder to generate ghost-free HDR images.
In the encoder, a context aggregation network and non-local attention block are adopted to optimize multi-scale features.
The decoder based on Swin Transformer is utilized to improve the reconstruction capability of the proposed model.
arXiv Detail & Related papers (2022-12-01T15:43:32Z) - Ghost-free High Dynamic Range Imaging with Context-aware Transformer [45.255802070953266]
We propose a novel Context-Aware Vision Transformer (CA-ViT) for ghost-free high dynamic range imaging.
The CA-ViT is designed as a dual-branch architecture, which can jointly capture both global and local dependencies.
By incorporating the CA-ViT as basic components, we further build the HDR-Transformer, a hierarchical network to reconstruct high-quality ghost-free HDR images.
arXiv Detail & Related papers (2022-08-10T03:00:10Z) - A Two-stage Deep Network for High Dynamic Range Image Reconstruction [0.883717274344425]
This study tackles the challenges of single-shot LDR to HDR mapping by proposing a novel two-stage deep network.
Notably, our proposed method aims to reconstruct an HDR image without knowing hardware information, including camera response function (CRF) and exposure settings.
arXiv Detail & Related papers (2021-04-19T15:19:17Z) - HDR-GAN: HDR Image Reconstruction from Multi-Exposed LDR Images with
Large Motions [62.44802076971331]
We propose a novel GAN-based model, HDR-GAN, for synthesizing HDR images from multi-exposed LDR images.
By incorporating adversarial learning, our method is able to produce faithful information in the regions with missing content.
arXiv Detail & Related papers (2020-07-03T11:42:35Z) - Single-Image HDR Reconstruction by Learning to Reverse the Camera
Pipeline [100.5353614588565]
We propose to incorporate the domain knowledge of the LDR image formation pipeline into our model.
We model the HDRto-LDR image formation pipeline as the (1) dynamic range clipping, (2) non-linear mapping from a camera response function, and (3) quantization.
We demonstrate that the proposed method performs favorably against state-of-the-art single-image HDR reconstruction algorithms.
arXiv Detail & Related papers (2020-04-02T17:59:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.