A Novel Unified Model for Multi-exposure Stereo Coding Based on Low Rank
Tucker-ALS and 3D-HEVC
- URL: http://arxiv.org/abs/2104.04726v1
- Date: Sat, 10 Apr 2021 10:10:14 GMT
- Title: A Novel Unified Model for Multi-exposure Stereo Coding Based on Low Rank
Tucker-ALS and 3D-HEVC
- Authors: Mansi Sharma, Aditya Wadaskar
- Abstract summary: We propose an efficient scheme for coding multi-exposure stereo images based on a tensor low-rank approximation scheme.
The multi-exposure fusion can be realized to generate HDR stereo output at the decoder for increased realism and binocular 3D depth cues.
The encoding with 3D-HEVC enhance the proposed scheme efficiency by exploiting intra-frame, inter-view and the inter-component redundancies in lowrank approximated representation.
- Score: 0.6091702876917279
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Display technology must offer high dynamic range (HDR) contrast-based depth
induction and 3D personalization simultaneously. Efficient algorithms to
compress HDR stereo data is critical. Direct capturing of HDR content is
complicated due to the high expense and scarcity of HDR cameras. The HDR 3D
images could be generated in low-cost by fusing low-dynamic-range (LDR) images
acquired using a stereo camera with various exposure settings. In this paper,
an efficient scheme for coding multi-exposure stereo images is proposed based
on a tensor low-rank approximation scheme. The multi-exposure fusion can be
realized to generate HDR stereo output at the decoder for increased realism and
exaggerated binocular 3D depth cues.
For exploiting spatial redundancy in LDR stereo images, the stack of
multi-exposure stereo images is decomposed into a set of projection matrices
and a core tensor following an alternating least squares Tucker decomposition
model. The compact, low-rank representation of the scene, thus, generated is
further processed by 3D extension of High Efficiency Video Coding standard. The
encoding with 3D-HEVC enhance the proposed scheme efficiency by exploiting
intra-frame, inter-view and the inter-component redundancies in low-rank
approximated representation. We consider constant luminance property of IPT and
Y'CbCr color space to precisely approximate intensity prediction and
perceptually minimize the encoding distortion. Besides, the proposed scheme
gives flexibility to adjust the bitrate of tensor latent components by changing
the rank of core tensor and its quantization. Extensive experiments on natural
scenes demonstrate that the proposed scheme outperforms state-of-the-art
JPEG-XT and 3D-HEVC range coding standards.
Related papers
- Pixel-Aligned Multi-View Generation with Depth Guided Decoder [86.1813201212539]
We propose a novel method for pixel-level image-to-multi-view generation.
Unlike prior work, we incorporate attention layers across multi-view images in the VAE decoder of a latent video diffusion model.
Our model enables better pixel alignment across multi-view images.
arXiv Detail & Related papers (2024-08-26T04:56:41Z) - HDRGS: High Dynamic Range Gaussian Splatting [19.119572715951172]
High Dynamic Range (GS) method enhances color dimensionality by luminance and uses an asymmetric grid for tone-mapping.
Our method surpasses current state-of-the-art techniques in both synthetic and real-world scenarios.
arXiv Detail & Related papers (2024-08-13T00:32:36Z) - Generating Content for HDR Deghosting from Frequency View [56.103761824603644]
Recent Diffusion Models (DMs) have been introduced in HDR imaging field.
DMs require extensive iterations with large models to estimate entire images.
We propose the Low-Frequency aware Diffusion (LF-Diff) model for ghost-free HDR imaging.
arXiv Detail & Related papers (2024-04-01T01:32:11Z) - Event-based Asynchronous HDR Imaging by Temporal Incident Light Modulation [54.64335350932855]
We propose a Pixel-Asynchronous HDR imaging system, based on key insights into the challenges in HDR imaging.
Our proposed Asyn system integrates the Dynamic Vision Sensors (DVS) with a set of LCD panels.
The LCD panels modulate the irradiance incident upon the DVS by altering their transparency, thereby triggering the pixel-independent event streams.
arXiv Detail & Related papers (2024-03-14T13:45:09Z) - Fast High Dynamic Range Radiance Fields for Dynamic Scenes [39.3304365600248]
We propose a dynamic HDR NeRF framework, named HDR-HexPlane, which can learn 3D scenes from dynamic 2D images captured with various exposures.
With the proposed model, high-quality novel-view images at any time point can be rendered with any desired exposure.
arXiv Detail & Related papers (2024-01-11T17:15:16Z) - LAN-HDR: Luminance-based Alignment Network for High Dynamic Range Video
Reconstruction [20.911738532410766]
We propose an end-to-end HDR video composition framework, which aligns LDR frames in feature space and then merges aligned features into an HDR frame.
In training, we adopt a temporal loss, in addition to frame reconstruction losses, to enhance temporal consistency and thus reduce flickering.
arXiv Detail & Related papers (2023-08-22T01:43:00Z) - Spatiotemporally Consistent HDR Indoor Lighting Estimation [66.26786775252592]
We propose a physically-motivated deep learning framework to solve the indoor lighting estimation problem.
Given a single LDR image with a depth map, our method predicts spatially consistent lighting at any given image position.
Our framework achieves photorealistic lighting prediction with higher quality compared to state-of-the-art single-image or video-based methods.
arXiv Detail & Related papers (2023-05-07T20:36:29Z) - Ghost-free High Dynamic Range Imaging via Hybrid CNN-Transformer and
Structure Tensor [12.167049432063132]
We present a hybrid model consisting of a convolutional encoder and a Transformer decoder to generate ghost-free HDR images.
In the encoder, a context aggregation network and non-local attention block are adopted to optimize multi-scale features.
The decoder based on Swin Transformer is utilized to improve the reconstruction capability of the proposed model.
arXiv Detail & Related papers (2022-12-01T15:43:32Z) - Deep Parametric 3D Filters for Joint Video Denoising and Illumination
Enhancement in Video Super Resolution [96.89588203312451]
This paper presents a new parametric representation called Deep Parametric 3D Filters (DP3DF)
DP3DF incorporates local information to enable simultaneous denoising, illumination enhancement, and SR efficiently in a single encoder-and-decoder network.
Also, a dynamic residual frame is jointly learned with the DP3DF via a shared backbone to further boost the SR quality.
arXiv Detail & Related papers (2022-07-05T03:57:25Z) - MEStereo-Du2CNN: A Novel Dual Channel CNN for Learning Robust Depth
Estimates from Multi-exposure Stereo Images for HDR 3D Applications [0.22940141855172028]
We develop a novel deep architecture for multi-exposure stereo depth estimation.
For the stereo depth estimation component of our architecture, a mono-to-stereo transfer learning approach is deployed.
In terms of performance, the proposed model surpasses state-of-the-art monocular and stereo depth estimation methods.
arXiv Detail & Related papers (2022-06-21T13:23:22Z) - Single-Image HDR Reconstruction by Learning to Reverse the Camera
Pipeline [100.5353614588565]
We propose to incorporate the domain knowledge of the LDR image formation pipeline into our model.
We model the HDRto-LDR image formation pipeline as the (1) dynamic range clipping, (2) non-linear mapping from a camera response function, and (3) quantization.
We demonstrate that the proposed method performs favorably against state-of-the-art single-image HDR reconstruction algorithms.
arXiv Detail & Related papers (2020-04-02T17:59:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.