Intrinsic Image Decomposition using Paradigms
- URL: http://arxiv.org/abs/2011.10512v1
- Date: Fri, 20 Nov 2020 17:10:12 GMT
- Title: Intrinsic Image Decomposition using Paradigms
- Authors: D. A. Forsyth and Jason J. Rock
- Abstract summary: Best modern image methods learn a map from image to albedo using rendered models and human judgements.
This paper describes a method that learns intrinsic image decomposition without seeing W annotations, rendered data, or ground truth data.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Intrinsic image decomposition is the classical task of mapping image to
albedo. The WHDR dataset allows methods to be evaluated by comparing
predictions to human judgements ("lighter", "same as", "darker"). The best
modern intrinsic image methods learn a map from image to albedo using rendered
models and human judgements. This is convenient for practical methods, but
cannot explain how a visual agent without geometric, surface and illumination
models and a renderer could learn to recover intrinsic images.
This paper describes a method that learns intrinsic image decomposition
without seeing WHDR annotations, rendered data, or ground truth data. The
method relies on paradigms - fake albedos and fake shading fields - together
with a novel smoothing procedure that ensures good behavior at short scales on
real images. Long scale error is controlled by averaging. Our method achieves
WHDR scores competitive with those of strong recent methods allowed to see
training WHDR annotations, rendered data, and ground truth data. Because our
method is unsupervised, we can compute estimates of the test/train variance of
WHDR scores; these are quite large, and it is unsafe to rely small differences
in reported WHDR.
Related papers
- Latent Intrinsics Emerge from Training to Relight [21.766083733177652]
This paper describes a relighting method that is entirely data-driven, where intrinsics and lighting are each represented as latent variables.
We show that albedo can be recovered from our latent intrinsics without using any example albedos, and that the albedos recovered are competitive with SOTA methods.
arXiv Detail & Related papers (2024-05-31T17:59:12Z) - Learning to Rank Patches for Unbiased Image Redundancy Reduction [80.93989115541966]
Images suffer from heavy spatial redundancy because pixels in neighboring regions are spatially correlated.
Existing approaches strive to overcome this limitation by reducing less meaningful image regions.
We propose a self-supervised framework for image redundancy reduction called Learning to Rank Patches.
arXiv Detail & Related papers (2024-03-31T13:12:41Z) - Detecting Generated Images by Real Images Only [64.12501227493765]
Existing generated image detection methods detect visual artifacts in generated images or learn discriminative features from both real and generated images by massive training.
This paper approaches the generated image detection problem from a new perspective: Start from real images.
By finding the commonality of real images and mapping them to a dense subspace in feature space, the goal is that generated images, regardless of their generative model, are then projected outside the subspace.
arXiv Detail & Related papers (2023-11-02T03:09:37Z) - Annotation Cost Efficient Active Learning for Content Based Image
Retrieval [1.6624384368855525]
We present an annotation cost efficient active learning (AL) method (denoted as ANNEAL)
The proposed method aims to iteratively enrich the training set by annotating the most informative image pairs as similar or dissimilar.
The code of ANNEAL is publicly available at https://git.tu-berlin.de/rsim/ANNEAL.
arXiv Detail & Related papers (2023-06-20T15:33:24Z) - Region-aware Attention for Image Inpainting [33.22497212024083]
We propose a novel region-aware attention (RA) module for inpainting images.
By avoiding the directly calculating corralation between each pixel pair in a single samples, the misleading of invalid information in holes can be avoided.
A learnable region dictionary (LRD) is introduced to store important information in the entire dataset.
Our methodscan generate semantically plausible results with realistic details.
arXiv Detail & Related papers (2022-04-03T06:26:22Z) - Perspective Flow Aggregation for Data-Limited 6D Object Pose Estimation [121.02948087956955]
For some applications, such as those in space or deep under water, acquiring real images, even unannotated, is virtually impossible.
We propose a method that can be trained solely on synthetic images, or optionally using a few additional real images.
It performs on par with methods that require annotated real images for training when not using any, and outperforms them considerably when using as few as twenty real images.
arXiv Detail & Related papers (2022-03-18T10:20:21Z) - SIRfyN: Single Image Relighting from your Neighbors [14.601975066158394]
We show how to relight a scene depicted in a single image, such that (a) the overall shading has changed and (b) the resulting image looks like a natural image of that scene.
arXiv Detail & Related papers (2021-12-08T17:05:57Z) - Contrastive Feature Loss for Image Prediction [55.373404869092866]
Training supervised image synthesis models requires a critic to compare two images: the ground truth to the result.
We introduce an information theory based approach to measuring similarity between two images.
We show that our formulation boosts the perceptual realism of output images when used as a drop-in replacement for the L1 loss.
arXiv Detail & Related papers (2021-11-12T20:39:52Z) - Inverse Problems Leveraging Pre-trained Contrastive Representations [88.70821497369785]
We study a new family of inverse problems for recovering representations of corrupted data.
We propose a supervised inversion method that uses a contrastive objective to obtain excellent representations for highly corrupted images.
Our method outperforms end-to-end baselines even with a fraction of the labeled data in a wide range of forward operators.
arXiv Detail & Related papers (2021-10-14T15:06:30Z) - Comparing to Learn: Surpassing ImageNet Pretraining on Radiographs By
Comparing Image Representations [39.08296644280442]
We propose a new pretraining method which learns from 700k radiographs given no manual annotations.
We call our method as Comparing to Learn (C2L) because it learns robust features by comparing different image representations.
The experimental results on radiographs show that C2L can outperform ImageNet pretraining and previous state-of-the-art approaches significantly.
arXiv Detail & Related papers (2020-07-15T01:14:34Z) - Syn2Real Transfer Learning for Image Deraining using Gaussian Processes [92.15895515035795]
CNN-based methods for image deraining have achieved excellent performance in terms of reconstruction error as well as visual quality.
Due to challenges in obtaining real world fully-labeled image deraining datasets, existing methods are trained only on synthetically generated data.
We propose a Gaussian Process-based semi-supervised learning framework which enables the network in learning to derain using synthetic dataset.
arXiv Detail & Related papers (2020-06-10T00:33:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.