Ghost-free High Dynamic Range Imaging with Context-aware Transformer
- URL: http://arxiv.org/abs/2208.05114v1
- Date: Wed, 10 Aug 2022 03:00:10 GMT
- Title: Ghost-free High Dynamic Range Imaging with Context-aware Transformer
- Authors: Zhen Liu, Yinglong Wang, Bing Zeng, Shuaicheng Liu
- Abstract summary: We propose a novel Context-Aware Vision Transformer (CA-ViT) for ghost-free high dynamic range imaging.
The CA-ViT is designed as a dual-branch architecture, which can jointly capture both global and local dependencies.
By incorporating the CA-ViT as basic components, we further build the HDR-Transformer, a hierarchical network to reconstruct high-quality ghost-free HDR images.
- Score: 45.255802070953266
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: High dynamic range (HDR) deghosting algorithms aim to generate ghost-free HDR
images with realistic details. Restricted by the locality of the receptive
field, existing CNN-based methods are typically prone to producing ghosting
artifacts and intensity distortions in the presence of large motion and severe
saturation. In this paper, we propose a novel Context-Aware Vision Transformer
(CA-ViT) for ghost-free high dynamic range imaging. The CA-ViT is designed as a
dual-branch architecture, which can jointly capture both global and local
dependencies. Specifically, the global branch employs a window-based
Transformer encoder to model long-range object movements and intensity
variations to solve ghosting. For the local branch, we design a local context
extractor (LCE) to capture short-range image features and use the channel
attention mechanism to select informative local details across the extracted
features to complement the global branch. By incorporating the CA-ViT as basic
components, we further build the HDR-Transformer, a hierarchical network to
reconstruct high-quality ghost-free HDR images. Extensive experiments on three
benchmark datasets show that our approach outperforms state-of-the-art methods
qualitatively and quantitatively with considerably reduced computational
budgets. Codes are available at
https://github.com/megvii-research/HDR-Transformer
Related papers
- TransY-Net:Learning Fully Transformer Networks for Change Detection of
Remote Sensing Images [64.63004710817239]
We propose a novel Transformer-based learning framework named TransY-Net for remote sensing image CD.
It improves the feature extraction from a global view and combines multi-level visual features in a pyramid manner.
Our proposed method achieves a new state-of-the-art performance on four optical and two SAR image CD benchmarks.
arXiv Detail & Related papers (2023-10-22T07:42:19Z) - IFT: Image Fusion Transformer for Ghost-free High Dynamic Range Imaging [26.127764855477782]
Multi-frame high dynamic range ( HDR) imaging aims to reconstruct ghost-free images with photo-realistic details from content-complementary but spatially misaligned low dynamic range (LDR) images.
Existing HDR algorithms are prone to producing ghosting artifacts as their methods fail to capture long-range dependencies between LDR frames with large motion in dynamic scenes.
We propose a novel image fusion transformer, referred to as IFT, which presents a fast global patch searching (FGPS) module followed by a self-cross fusion module (SCF) for ghost-free HDR imaging.
arXiv Detail & Related papers (2023-09-26T15:38:52Z) - Alignment-free HDR Deghosting with Semantics Consistent Transformer [76.91669741684173]
High dynamic range imaging aims to retrieve information from multiple low-dynamic range inputs to generate realistic output.
Existing methods often focus on the spatial misalignment across input frames caused by the foreground and/or camera motion.
We propose a novel alignment-free network with a Semantics Consistent Transformer (SCTNet) with both spatial and channel attention modules.
arXiv Detail & Related papers (2023-05-29T15:03:23Z) - A Unified HDR Imaging Method with Pixel and Patch Level [41.14378863436963]
We propose a hybrid HDR deghosting network, called HyNet, to generate visually pleasing HDR images.
Experiments demonstrate that HyNet outperforms state-of-the-art methods both quantitatively and qualitatively, achieving appealing HDR visualization with unified textures and colors.
arXiv Detail & Related papers (2023-04-14T06:21:57Z) - High Dynamic Range Imaging with Context-aware Transformer [3.1892103878735454]
We propose a novel hierarchical dual Transformer (HDT) method for ghost-free images.
First, we use a CNN-based head with spatial attention mechanisms to extract features from all the LDR images.
Second, the LDR features are delivered to the Transformer, while the local details are extracted using the channel attention mechanism.
arXiv Detail & Related papers (2023-04-10T06:56:01Z) - DLGSANet: Lightweight Dynamic Local and Global Self-Attention Networks
for Image Super-Resolution [83.47467223117361]
We propose an effective lightweight dynamic local and global self-attention network (DLGSANet) to solve image super-resolution.
Motivated by the network designs of Transformers, we develop a simple yet effective multi-head dynamic local self-attention (MHDLSA) module to extract local features efficiently.
To overcome this problem, we develop a sparse global self-attention (SparseGSA) module to select the most useful similarity values.
arXiv Detail & Related papers (2023-01-05T12:06:47Z) - Ghost-free High Dynamic Range Imaging via Hybrid CNN-Transformer and
Structure Tensor [12.167049432063132]
We present a hybrid model consisting of a convolutional encoder and a Transformer decoder to generate ghost-free HDR images.
In the encoder, a context aggregation network and non-local attention block are adopted to optimize multi-scale features.
The decoder based on Swin Transformer is utilized to improve the reconstruction capability of the proposed model.
arXiv Detail & Related papers (2022-12-01T15:43:32Z) - DynaST: Dynamic Sparse Transformer for Exemplar-Guided Image Generation [56.514462874501675]
We propose a dynamic sparse attention based Transformer model to achieve fine-level matching with favorable efficiency.
The heart of our approach is a novel dynamic-attention unit, dedicated to covering the variation on the optimal number of tokens one position should focus on.
Experiments on three applications, pose-guided person image generation, edge-based face synthesis, and undistorted image style transfer, demonstrate that DynaST achieves superior performance in local details.
arXiv Detail & Related papers (2022-07-13T11:12:03Z) - Wavelet-Based Network For High Dynamic Range Imaging [64.66969585951207]
Existing methods, such as optical flow based and end-to-end deep learning based solutions, are error-prone either in detail restoration or ghosting artifacts removal.
In this work, we propose a novel frequency-guided end-to-end deep neural network (FNet) to conduct HDR fusion in the frequency domain, and Wavelet Transform (DWT) is used to decompose inputs into different frequency bands.
The low-frequency signals are used to avoid specific ghosting artifacts, while the high-frequency signals are used for preserving details.
arXiv Detail & Related papers (2021-08-03T12:26:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.