LFSRDiff: Light Field Image Super-Resolution via Diffusion Models
- URL: http://arxiv.org/abs/2311.16517v1
- Date: Mon, 27 Nov 2023 07:31:12 GMT
- Title: LFSRDiff: Light Field Image Super-Resolution via Diffusion Models
- Authors: Wentao Chao, Fuqing Duan, Xuechun Wang, Yingqian Wang, Guanghui Wang
- Abstract summary: Light field (LF) image super-resolution (SR) is a challenging problem due to its inherent ill-posed nature.
mainstream LF image SR methods typically adopt a deterministic approach, generating only a single output supervised by pixel-wise loss functions.
We introduce LFSRDiff, the first diffusion-based LF image SR model, by incorporating the LF disentanglement mechanism.
- Score: 18.20217829625834
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Light field (LF) image super-resolution (SR) is a challenging problem due to
its inherent ill-posed nature, where a single low-resolution (LR) input LF
image can correspond to multiple potential super-resolved outcomes. Despite
this complexity, mainstream LF image SR methods typically adopt a deterministic
approach, generating only a single output supervised by pixel-wise loss
functions. This tendency often results in blurry and unrealistic results.
Although diffusion models can capture the distribution of potential SR results
by iteratively predicting Gaussian noise during the denoising process, they are
primarily designed for general images and struggle to effectively handle the
unique characteristics and information present in LF images. To address these
limitations, we introduce LFSRDiff, the first diffusion-based LF image SR
model, by incorporating the LF disentanglement mechanism. Our novel
contribution includes the introduction of a disentangled U-Net for diffusion
models, enabling more effective extraction and fusion of both spatial and
angular information within LF images. Through comprehensive experimental
evaluations and comparisons with the state-of-the-art LF image SR methods, the
proposed approach consistently produces diverse and realistic SR results. It
achieves the highest perceptual metric in terms of LPIPS. It also demonstrates
the ability to effectively control the trade-off between perception and
distortion. The code is available at
\url{https://github.com/chaowentao/LFSRDiff}.
Related papers
- LoLiSRFlow: Joint Single Image Low-light Enhancement and
Super-resolution via Cross-scale Transformer-based Conditional Flow [8.929704596997913]
We propose a normalizing flow network, dubbed LoLiSRFLow, to consider the degradation mechanism inherent in Low-Light Enhancement (LLE) and Super- Resolution (SR)
LoLiSRFLow learns the conditional probability distribution over a variety of feasible solutions for high-resolution well-exposed images.
We also propose a synthetic dataset modeling the realistic low-light low-resolution degradation, named DFSR-LLE, containing 7100 low-resolution dark-light/high-resolution normal sharp pairs.
arXiv Detail & Related papers (2024-02-29T05:40:43Z) - Diffusion-based Light Field Synthesis [50.24624071354433]
LFdiff is a diffusion-based generative framework tailored for LF synthesis.
We propose DistgUnet, a disentanglement-based noise estimation network, to harness comprehensive LF representations.
Extensive experiments demonstrate that LFdiff excels in synthesizing visually pleasing and disparity-controllable light fields.
arXiv Detail & Related papers (2024-02-01T13:13:16Z) - Local Implicit Normalizing Flow for Arbitrary-Scale Image
Super-Resolution [11.653044501483667]
We propose "Local Implicit Normalizing Flow" (LINF) as a unified solution to the above problems.
LINF models the distribution of texture details under different scaling factors with normalizing flow.
We show that LINF achieves the state-of-the-art perceptual quality compared with prior arbitrary-scale SR methods.
arXiv Detail & Related papers (2023-03-09T10:20:07Z) - Blind Super-Resolution for Remote Sensing Images via Conditional
Stochastic Normalizing Flows [14.882417028542855]
We propose a novel blind SR framework based on the normalizing flow (BlindSRSNF) to address the above problems.
BlindSRSNF learns the conditional probability distribution over the high-resolution image space given a low-resolution (LR) image by explicitly optimizing the variational bound on the likelihood.
We show that the proposed algorithm can obtain SR results with excellent visual perception quality on both simulated LR and real-world RSIs.
arXiv Detail & Related papers (2022-10-14T12:37:32Z) - Real-World Light Field Image Super-Resolution via Degradation Modulation [59.68036846233918]
We propose a simple yet effective method for real-world LF image SR.
A practical LF degradation model is developed to formulate the degradation process of real LF images.
A convolutional neural network is designed to incorporate the degradation prior to the SR process.
arXiv Detail & Related papers (2022-06-13T14:44:46Z) - SRDiff: Single Image Super-Resolution with Diffusion Probabilistic
Models [19.17571465274627]
Single image super-resolution (SISR) aims to reconstruct high-resolution (HR) images from the given low-resolution (LR) ones.
We propose a novel single image super-resolution diffusion probabilistic model (SRDiff)
SRDiff is optimized with a variant of the variational bound on the data likelihood and can provide diverse and realistic SR predictions.
arXiv Detail & Related papers (2021-04-30T12:31:25Z) - Frequency Consistent Adaptation for Real World Super Resolution [64.91914552787668]
We propose a novel Frequency Consistent Adaptation (FCA) that ensures the frequency domain consistency when applying Super-Resolution (SR) methods to the real scene.
We estimate degradation kernels from unsupervised images and generate the corresponding Low-Resolution (LR) images.
Based on the domain-consistent LR-HR pairs, we train easy-implemented Convolutional Neural Network (CNN) SR models.
arXiv Detail & Related papers (2020-12-18T08:25:39Z) - Deep Selective Combinatorial Embedding and Consistency Regularization
for Light Field Super-resolution [93.95828097088608]
Light field (LF) images acquired by hand-held devices usually suffer from low spatial resolution.
The high-dimensionality characteristic and complex geometrical structure of LF images make the problem more challenging than traditional single-image SR.
We propose a novel learning-based LF spatial SR framework to explore the coherence among LF sub-aperture images.
Experimental results over both synthetic and real-world LF datasets demonstrate the significant advantage of our approach over state-of-the-art methods.
arXiv Detail & Related papers (2020-09-26T08:34:37Z) - SRFlow: Learning the Super-Resolution Space with Normalizing Flow [176.07982398988747]
Super-resolution is an ill-posed problem, since it allows for multiple predictions for a given low-resolution image.
We propose SRFlow: a normalizing flow based super-resolution method capable of learning the conditional distribution of the output.
Our model is trained in a principled manner using a single loss, namely the negative log-likelihood.
arXiv Detail & Related papers (2020-06-25T06:34:04Z) - Light Field Spatial Super-resolution via Deep Combinatorial Geometry
Embedding and Structural Consistency Regularization [99.96632216070718]
Light field (LF) images acquired by hand-held devices usually suffer from low spatial resolution.
The high-dimensional spatiality characteristic and complex geometrical structure of LF images make the problem more challenging than traditional single-image SR.
We propose a novel learning-based LF framework, in which each view of an LF image is first individually super-resolved.
arXiv Detail & Related papers (2020-04-05T14:39:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.