Ghost-free High Dynamic Range Imaging via Hybrid CNN-Transformer and
Structure Tensor
- URL: http://arxiv.org/abs/2212.00595v1
- Date: Thu, 1 Dec 2022 15:43:32 GMT
- Title: Ghost-free High Dynamic Range Imaging via Hybrid CNN-Transformer and
Structure Tensor
- Authors: Yu Yuan and Jiaqi Wu and Zhongliang Jing and Henry Leung and Han Pan
- Abstract summary: We present a hybrid model consisting of a convolutional encoder and a Transformer decoder to generate ghost-free HDR images.
In the encoder, a context aggregation network and non-local attention block are adopted to optimize multi-scale features.
The decoder based on Swin Transformer is utilized to improve the reconstruction capability of the proposed model.
- Score: 12.167049432063132
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Eliminating ghosting artifacts due to moving objects is a challenging problem
in high dynamic range (HDR) imaging. In this letter, we present a hybrid model
consisting of a convolutional encoder and a Transformer decoder to generate
ghost-free HDR images. In the encoder, a context aggregation network and
non-local attention block are adopted to optimize multi-scale features and
capture both global and local dependencies of multiple low dynamic range (LDR)
images. The decoder based on Swin Transformer is utilized to improve the
reconstruction capability of the proposed model. Motivated by the phenomenal
difference between the presence and absence of artifacts under the field of
structure tensor (ST), we integrate the ST information of LDR images as
auxiliary inputs of the network and use ST loss to further constrain artifacts.
Different from previous approaches, our network is capable of processing an
arbitrary number of input LDR images. Qualitative and quantitative experiments
demonstrate the effectiveness of the proposed method by comparing it with
existing state-of-the-art HDR deghosting models. Codes are available at
https://github.com/pandayuanyu/HSTHdr.
Related papers
- HDRTransDC: High Dynamic Range Image Reconstruction with Transformer Deformation Convolution [21.870772317331447]
High Dynamic Range (CAM) imaging aims to generate an artifact-free HDR image with realistic details by fusing multi-exposure Low Dynamic Range (LDR) images.
For the purpose of eliminating fusion distortions, we propose DWFB to spatially adaptively select useful information across frames.
arXiv Detail & Related papers (2024-03-11T15:48:17Z) - Towards High-quality HDR Deghosting with Conditional Diffusion Models [88.83729417524823]
High Dynamic Range (LDR) images can be recovered from several Low Dynamic Range (LDR) images by existing Deep Neural Networks (DNNs) techniques.
DNNs still generate ghosting artifacts when LDR images have saturation and large motion.
We formulate the HDR deghosting problem as an image generation that leverages LDR features as the diffusion model's condition.
arXiv Detail & Related papers (2023-11-02T01:53:55Z) - Distance Weighted Trans Network for Image Completion [52.318730994423106]
We propose a new architecture that relies on Distance-based Weighted Transformer (DWT) to better understand the relationships between an image's components.
CNNs are used to augment the local texture information of coarse priors.
DWT blocks are used to recover certain coarse textures and coherent visual structures.
arXiv Detail & Related papers (2023-10-11T12:46:11Z) - Alignment-free HDR Deghosting with Semantics Consistent Transformer [76.91669741684173]
High dynamic range imaging aims to retrieve information from multiple low-dynamic range inputs to generate realistic output.
Existing methods often focus on the spatial misalignment across input frames caused by the foreground and/or camera motion.
We propose a novel alignment-free network with a Semantics Consistent Transformer (SCTNet) with both spatial and channel attention modules.
arXiv Detail & Related papers (2023-05-29T15:03:23Z) - A Unified HDR Imaging Method with Pixel and Patch Level [41.14378863436963]
We propose a hybrid HDR deghosting network, called HyNet, to generate visually pleasing HDR images.
Experiments demonstrate that HyNet outperforms state-of-the-art methods both quantitatively and qualitatively, achieving appealing HDR visualization with unified textures and colors.
arXiv Detail & Related papers (2023-04-14T06:21:57Z) - High Dynamic Range Imaging with Context-aware Transformer [3.1892103878735454]
We propose a novel hierarchical dual Transformer (HDT) method for ghost-free images.
First, we use a CNN-based head with spatial attention mechanisms to extract features from all the LDR images.
Second, the LDR features are delivered to the Transformer, while the local details are extracted using the channel attention mechanism.
arXiv Detail & Related papers (2023-04-10T06:56:01Z) - Ghost-free High Dynamic Range Imaging with Context-aware Transformer [45.255802070953266]
We propose a novel Context-Aware Vision Transformer (CA-ViT) for ghost-free high dynamic range imaging.
The CA-ViT is designed as a dual-branch architecture, which can jointly capture both global and local dependencies.
By incorporating the CA-ViT as basic components, we further build the HDR-Transformer, a hierarchical network to reconstruct high-quality ghost-free HDR images.
arXiv Detail & Related papers (2022-08-10T03:00:10Z) - Wavelet-Based Network For High Dynamic Range Imaging [64.66969585951207]
Existing methods, such as optical flow based and end-to-end deep learning based solutions, are error-prone either in detail restoration or ghosting artifacts removal.
In this work, we propose a novel frequency-guided end-to-end deep neural network (FNet) to conduct HDR fusion in the frequency domain, and Wavelet Transform (DWT) is used to decompose inputs into different frequency bands.
The low-frequency signals are used to avoid specific ghosting artifacts, while the high-frequency signals are used for preserving details.
arXiv Detail & Related papers (2021-08-03T12:26:33Z) - HDRUNet: Single Image HDR Reconstruction with Denoising and
Dequantization [39.82945546614887]
We propose a novel learning-based approach using a spatially dynamic encoder-decoder network, HDRUNet, to learn an end-to-end mapping for single image HDR reconstruction.
Our method achieves the state-of-the-art performance in quantitative comparisons and visual quality.
arXiv Detail & Related papers (2021-05-27T12:12:34Z) - Single-Image HDR Reconstruction by Learning to Reverse the Camera
Pipeline [100.5353614588565]
We propose to incorporate the domain knowledge of the LDR image formation pipeline into our model.
We model the HDRto-LDR image formation pipeline as the (1) dynamic range clipping, (2) non-linear mapping from a camera response function, and (3) quantization.
We demonstrate that the proposed method performs favorably against state-of-the-art single-image HDR reconstruction algorithms.
arXiv Detail & Related papers (2020-04-02T17:59:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.