Ultra-High-Definition Dynamic Multi-Exposure Image Fusion via Infinite Pixel Learning
- URL: http://arxiv.org/abs/2412.11685v1
- Date: Mon, 16 Dec 2024 11:55:26 GMT
- Title: Ultra-High-Definition Dynamic Multi-Exposure Image Fusion via Infinite Pixel Learning
- Authors: Xingchi Chen, Zhuoran Zheng, Xuerui Li, Yuying Chen, Shu Wang, Wenqi Ren,
- Abstract summary: Existing methods for fusing multi-exposure images in dynamic scenes are designed for low-resolution images.
We propose a novel learning paradigm to achieve UHD multi-exposure dynamic scene image fusion on a single consumer-grade GPU.
Our method maintains high-quality visual performance while fusing UHD dynamic multi-exposure images in real-time.
- Score: 36.536727716387446
- License:
- Abstract: With the continuous improvement of device imaging resolution, the popularity of Ultra-High-Definition (UHD) images is increasing. Unfortunately, existing methods for fusing multi-exposure images in dynamic scenes are designed for low-resolution images, which makes them inefficient for generating high-quality UHD images on a resource-constrained device. To alleviate the limitations of extremely long-sequence inputs, inspired by the Large Language Model (LLM) for processing infinitely long texts, we propose a novel learning paradigm to achieve UHD multi-exposure dynamic scene image fusion on a single consumer-grade GPU, named Infinite Pixel Learning (IPL). The design of our approach comes from three key components: The first step is to slice the input sequences to relieve the pressure generated by the model processing the data stream; Second, we develop an attention cache technique, which is similar to KV cache for infinite data stream processing; Finally, we design a method for attention cache compression to alleviate the storage burden of the cache on the device. In addition, we provide a new UHD benchmark to evaluate the effectiveness of our method. Extensive experimental results show that our method maintains high-quality visual performance while fusing UHD dynamic multi-exposure images in real-time (>40fps) on a single consumer-grade GPU.
Related papers
- Dropout the High-rate Downsampling: A Novel Design Paradigm for UHD Image Restoration [11.866565346920781]
D2Net enables direct full-resolution inference on Ultra-high-definition (UHD) images.
We show that our model achieves better quantitative and qualitative results than state-of-the-art methods.
arXiv Detail & Related papers (2024-11-10T13:05:36Z) - Pixel-Aligned Multi-View Generation with Depth Guided Decoder [86.1813201212539]
We propose a novel method for pixel-level image-to-multi-view generation.
Unlike prior work, we incorporate attention layers across multi-view images in the VAE decoder of a latent video diffusion model.
Our model enables better pixel alignment across multi-view images.
arXiv Detail & Related papers (2024-08-26T04:56:41Z) - Towards Ultra-High-Definition Image Deraining: A Benchmark and An Efficient Method [42.331058889312466]
This paper contributes the first large-scale UHD image deraining dataset, 4K-Rain13k, that contains 13,000 image pairs at 4K resolution.
We develop an effective and efficient vision-based architecture (UDR-Mixer) to better solve this task.
arXiv Detail & Related papers (2024-05-27T11:45:08Z) - HINT: High-quality INPainting Transformer with Mask-Aware Encoding and
Enhanced Attention [14.055584700641212]
Existing image inpainting methods leverage convolution-based downsampling approaches to reduce spatial dimensions.
We propose an end-to-end High-quality INpainting Transformer, abbreviated as HINT, which consists of a novel mask-aware pixel-shuffle downsampling module.
We demonstrate the superior performance of HINT compared to contemporary state-of-the-art models on four datasets.
arXiv Detail & Related papers (2024-02-22T00:14:26Z) - Gated Multi-Resolution Transfer Network for Burst Restoration and
Enhancement [75.25451566988565]
We propose a novel Gated Multi-Resolution Transfer Network (GMTNet) to reconstruct a spatially precise high-quality image from a burst of low-quality raw images.
Detailed experimental analysis on five datasets validates our approach and sets a state-of-the-art for burst super-resolution, burst denoising, and low-light burst enhancement.
arXiv Detail & Related papers (2023-04-13T17:54:00Z) - Embedding Fourier for Ultra-High-Definition Low-Light Image Enhancement [78.67036949708795]
Ultra-High-Definition (UHD) photo has gradually become the standard configuration in advanced imaging devices.
We propose a new solution, UHDFour, that embeds Fourier transform into a cascaded network.
We also contribute the first real UHD LLIE dataset, textbfUHD-LL, that contains 2,150 low-noise/normal-clear 4K image pairs.
arXiv Detail & Related papers (2023-02-23T07:43:41Z) - Deep Progressive Feature Aggregation Network for High Dynamic Range
Imaging [24.94466716276423]
We propose a deep progressive feature aggregation network for improving HDR imaging quality in dynamic scenes.
Our method implicitly samples high-correspondence features and aggregates them in a coarse-to-fine manner for alignment.
Experiments show that our proposed method can achieve state-of-the-art performance under different scenes.
arXiv Detail & Related papers (2022-08-04T04:37:35Z) - Learning Regularized Multi-Scale Feature Flow for High Dynamic Range
Imaging [29.691689596845112]
We propose a deep network that tries to learn multi-scale feature flow guided by the regularized loss.
It first extracts multi-scale features and then aligns features from non-reference images.
After alignment, we use residual channel attention blocks to merge the features from different images.
arXiv Detail & Related papers (2022-07-06T09:37:28Z) - DeepMultiCap: Performance Capture of Multiple Characters Using Sparse
Multiview Cameras [63.186486240525554]
DeepMultiCap is a novel method for multi-person performance capture using sparse multi-view cameras.
Our method can capture time varying surface details without the need of using pre-scanned template models.
arXiv Detail & Related papers (2021-05-01T14:32:13Z) - Early Exit or Not: Resource-Efficient Blind Quality Enhancement for
Compressed Images [54.40852143927333]
Lossy image compression is pervasively conducted to save communication bandwidth, resulting in undesirable compression artifacts.
We propose a resource-efficient blind quality enhancement (RBQE) approach for compressed images.
Our approach can automatically decide to terminate or continue enhancement according to the assessed quality of enhanced images.
arXiv Detail & Related papers (2020-06-30T07:38:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.