Intrinsic Image Decomposition via Ordinal Shading
- URL: http://arxiv.org/abs/2311.12792v1
- Date: Tue, 21 Nov 2023 18:58:01 GMT
- Title: Intrinsic Image Decomposition via Ordinal Shading
- Authors: Chris Careaga and Ya\u{g}{\i}z Aksoy
- Abstract summary: Intrinsic decomposition is a fundamental mid-level vision problem that plays a crucial role in inverse rendering and computational photography pipelines.
We present a dense ordinal shading formulation using a shift- and scale-invariant loss to estimate ordinal shading cues.
We then combine low- and high-resolution ordinal estimations using a second network to generate a shading estimate with both global coherency and local details.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Intrinsic decomposition is a fundamental mid-level vision problem that plays
a crucial role in various inverse rendering and computational photography
pipelines. Generating highly accurate intrinsic decompositions is an inherently
under-constrained task that requires precisely estimating continuous-valued
shading and albedo. In this work, we achieve high-resolution intrinsic
decomposition by breaking the problem into two parts. First, we present a dense
ordinal shading formulation using a shift- and scale-invariant loss in order to
estimate ordinal shading cues without restricting the predictions to obey the
intrinsic model. We then combine low- and high-resolution ordinal estimations
using a second network to generate a shading estimate with both global
coherency and local details. We encourage the model to learn an accurate
decomposition by computing losses on the estimated shading as well as the
albedo implied by the intrinsic model. We develop a straightforward method for
generating dense pseudo ground truth using our model's predictions and
multi-illumination data, enabling generalization to in-the-wild imagery. We
present an exhaustive qualitative and quantitative analysis of our predicted
intrinsic components against state-of-the-art methods. Finally, we demonstrate
the real-world applicability of our estimations by performing otherwise
difficult editing tasks such as recoloring and relighting.
Related papers
- Iteratively Refined Image Reconstruction with Learned Attentive Regularizers [14.93489065234423]
We propose a regularization scheme for image reconstruction that leverages the power of deep learning.
Our scheme is interpretable because it corresponds to the minimization of a series of convex problems.
We offer a promising balance between interpretability, theoretical guarantees, reliability, and performance.
arXiv Detail & Related papers (2024-07-09T07:22:48Z) - Exploiting Diffusion Prior for Generalizable Dense Prediction [85.4563592053464]
Recent advanced Text-to-Image (T2I) diffusion models are sometimes too imaginative for existing off-the-shelf dense predictors to estimate.
We introduce DMP, a pipeline utilizing pre-trained T2I models as a prior for dense prediction tasks.
Despite limited-domain training data, the approach yields faithful estimations for arbitrary images, surpassing existing state-of-the-art algorithms.
arXiv Detail & Related papers (2023-11-30T18:59:44Z) - The Perception-Robustness Tradeoff in Deterministic Image Restoration [34.50287066865267]
We study the behavior of deterministic methods for solving inverse problems in imaging.
To approach perfect quality and perfect consistency, the Lipschitz constant of the model must grow to infinity.
We demonstrate our theory on single image super-resolution algorithms, addressing both noisy and noiseless settings.
arXiv Detail & Related papers (2023-11-14T18:30:34Z) - BID-NeRF: RGB-D image pose estimation with inverted Neural Radiance
Fields [0.0]
We aim to improve the Inverted Neural Radiance Fields (iNeRF) algorithm which defines the image pose estimation problem as a NeRF based iterative linear optimization.
NeRFs are novel neural space representation models that can synthesize photorealistic novel views of real-world scenes or objects.
arXiv Detail & Related papers (2023-10-05T14:27:06Z) - Regularized Vector Quantization for Tokenized Image Synthesis [126.96880843754066]
Quantizing images into discrete representations has been a fundamental problem in unified generative modeling.
deterministic quantization suffers from severe codebook collapse and misalignment with inference stage while quantization suffers from low codebook utilization and reconstruction objective.
This paper presents a regularized vector quantization framework that allows to mitigate perturbed above issues effectively by applying regularization from two perspectives.
arXiv Detail & Related papers (2023-03-11T15:20:54Z) - Invertible Rescaling Network and Its Extensions [118.72015270085535]
In this work, we propose a novel invertible framework to model the bidirectional degradation and restoration from a new perspective.
We develop invertible models to generate valid degraded images and transform the distribution of lost contents.
Then restoration is made tractable by applying the inverse transformation on the generated degraded image together with a randomly-drawn latent variable.
arXiv Detail & Related papers (2022-10-09T06:58:58Z) - Uncertainty-Aware Adaptation for Self-Supervised 3D Human Pose
Estimation [70.32536356351706]
We introduce MRP-Net that constitutes a common deep network backbone with two output heads subscribing to two diverse configurations.
We derive suitable measures to quantify prediction uncertainty at both pose and joint level.
We present a comprehensive evaluation of the proposed approach and demonstrate state-of-the-art performance on benchmark datasets.
arXiv Detail & Related papers (2022-03-29T07:14:58Z) - RNNPose: Recurrent 6-DoF Object Pose Refinement with Robust
Correspondence Field Estimation and Pose Optimization [46.144194562841435]
We propose a framework based on a recurrent neural network (RNN) for object pose refinement.
The problem is formulated as a non-linear least squares problem based on the estimated correspondence field.
The correspondence field estimation and pose refinement are conducted alternatively in each iteration to recover accurate object poses.
arXiv Detail & Related papers (2022-03-24T06:24:55Z) - Deep Variational Network Toward Blind Image Restoration [60.45350399661175]
Blind image restoration is a common yet challenging problem in computer vision.
We propose a novel blind image restoration method, aiming to integrate both the advantages of them.
Experiments on two typical blind IR tasks, namely image denoising and super-resolution, demonstrate that the proposed method achieves superior performance over current state-of-the-arts.
arXiv Detail & Related papers (2020-08-25T03:30:53Z) - Consistency Guided Scene Flow Estimation [159.24395181068218]
CGSF is a self-supervised framework for the joint reconstruction of 3D scene structure and motion from stereo video.
We show that the proposed model can reliably predict disparity and scene flow in challenging imagery.
It achieves better generalization than the state-of-the-art, and adapts quickly and robustly to unseen domains.
arXiv Detail & Related papers (2020-06-19T17:28:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.