Intrinsic Image Decomposition via Ordinal Shading
- URL: http://arxiv.org/abs/2311.12792v1
- Date: Tue, 21 Nov 2023 18:58:01 GMT
- Title: Intrinsic Image Decomposition via Ordinal Shading
- Authors: Chris Careaga and Ya\u{g}{\i}z Aksoy
- Abstract summary: Intrinsic decomposition is a fundamental mid-level vision problem that plays a crucial role in inverse rendering and computational photography pipelines.
We present a dense ordinal shading formulation using a shift- and scale-invariant loss to estimate ordinal shading cues.
We then combine low- and high-resolution ordinal estimations using a second network to generate a shading estimate with both global coherency and local details.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Intrinsic decomposition is a fundamental mid-level vision problem that plays
a crucial role in various inverse rendering and computational photography
pipelines. Generating highly accurate intrinsic decompositions is an inherently
under-constrained task that requires precisely estimating continuous-valued
shading and albedo. In this work, we achieve high-resolution intrinsic
decomposition by breaking the problem into two parts. First, we present a dense
ordinal shading formulation using a shift- and scale-invariant loss in order to
estimate ordinal shading cues without restricting the predictions to obey the
intrinsic model. We then combine low- and high-resolution ordinal estimations
using a second network to generate a shading estimate with both global
coherency and local details. We encourage the model to learn an accurate
decomposition by computing losses on the estimated shading as well as the
albedo implied by the intrinsic model. We develop a straightforward method for
generating dense pseudo ground truth using our model's predictions and
multi-illumination data, enabling generalization to in-the-wild imagery. We
present an exhaustive qualitative and quantitative analysis of our predicted
intrinsic components against state-of-the-art methods. Finally, we demonstrate
the real-world applicability of our estimations by performing otherwise
difficult editing tasks such as recoloring and relighting.
Related papers
- InDeed: Interpretable image deep decomposition with guaranteed generalizability [28.595151003310452]
Image decomposition aims to analyze an image into elementary components.
Deep learning can be powerful for such tasks, but its combination with a focus on interpretability and generalizability is rarely explored.
We introduce a novel framework for interpretable deep image decomposition, combining hierarchical Bayesian modeling and deep learning.
arXiv Detail & Related papers (2025-01-02T07:58:26Z) - Orthogonal Subspace Decomposition for Generalizable AI-Generated Image Detection [58.87142367781417]
A naively trained detector tends to favor overfitting to the limited and monotonous fake patterns, causing the feature space to become highly constrained and low-ranked.
One potential remedy is incorporating the pre-trained knowledge within the vision foundation models to expand the feature space.
By freezing the principal components and adapting only the remained components, we preserve the pre-trained knowledge while learning forgery-related patterns.
arXiv Detail & Related papers (2024-11-23T19:10:32Z) - Iteratively Refined Image Reconstruction with Learned Attentive Regularizers [14.93489065234423]
We propose a regularization scheme for image reconstruction that leverages the power of deep learning.
Our scheme is interpretable because it corresponds to the minimization of a series of convex problems.
We offer a promising balance between interpretability, theoretical guarantees, reliability, and performance.
arXiv Detail & Related papers (2024-07-09T07:22:48Z) - Exploiting Diffusion Prior for Generalizable Dense Prediction [85.4563592053464]
Recent advanced Text-to-Image (T2I) diffusion models are sometimes too imaginative for existing off-the-shelf dense predictors to estimate.
We introduce DMP, a pipeline utilizing pre-trained T2I models as a prior for dense prediction tasks.
Despite limited-domain training data, the approach yields faithful estimations for arbitrary images, surpassing existing state-of-the-art algorithms.
arXiv Detail & Related papers (2023-11-30T18:59:44Z) - The Perception-Robustness Tradeoff in Deterministic Image Restoration [34.50287066865267]
We study the behavior of deterministic methods for solving inverse problems in imaging.
To approach perfect quality and perfect consistency, the Lipschitz constant of the model must grow to infinity.
We demonstrate our theory on single image super-resolution algorithms, addressing both noisy and noiseless settings.
arXiv Detail & Related papers (2023-11-14T18:30:34Z) - BID-NeRF: RGB-D image pose estimation with inverted Neural Radiance
Fields [0.0]
We aim to improve the Inverted Neural Radiance Fields (iNeRF) algorithm which defines the image pose estimation problem as a NeRF based iterative linear optimization.
NeRFs are novel neural space representation models that can synthesize photorealistic novel views of real-world scenes or objects.
arXiv Detail & Related papers (2023-10-05T14:27:06Z) - Invertible Rescaling Network and Its Extensions [118.72015270085535]
In this work, we propose a novel invertible framework to model the bidirectional degradation and restoration from a new perspective.
We develop invertible models to generate valid degraded images and transform the distribution of lost contents.
Then restoration is made tractable by applying the inverse transformation on the generated degraded image together with a randomly-drawn latent variable.
arXiv Detail & Related papers (2022-10-09T06:58:58Z) - Uncertainty-Aware Adaptation for Self-Supervised 3D Human Pose
Estimation [70.32536356351706]
We introduce MRP-Net that constitutes a common deep network backbone with two output heads subscribing to two diverse configurations.
We derive suitable measures to quantify prediction uncertainty at both pose and joint level.
We present a comprehensive evaluation of the proposed approach and demonstrate state-of-the-art performance on benchmark datasets.
arXiv Detail & Related papers (2022-03-29T07:14:58Z) - Deep Variational Network Toward Blind Image Restoration [60.45350399661175]
Blind image restoration is a common yet challenging problem in computer vision.
We propose a novel blind image restoration method, aiming to integrate both the advantages of them.
Experiments on two typical blind IR tasks, namely image denoising and super-resolution, demonstrate that the proposed method achieves superior performance over current state-of-the-arts.
arXiv Detail & Related papers (2020-08-25T03:30:53Z) - Consistency Guided Scene Flow Estimation [159.24395181068218]
CGSF is a self-supervised framework for the joint reconstruction of 3D scene structure and motion from stereo video.
We show that the proposed model can reliably predict disparity and scene flow in challenging imagery.
It achieves better generalization than the state-of-the-art, and adapts quickly and robustly to unseen domains.
arXiv Detail & Related papers (2020-06-19T17:28:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.