Relighting from a Single Image: Datasets and Deep Intrinsic-based Architecture
- URL: http://arxiv.org/abs/2409.18770v1
- Date: Fri, 27 Sep 2024 14:15:02 GMT
- Title: Relighting from a Single Image: Datasets and Deep Intrinsic-based Architecture
- Authors: Yixiong Yang, Hassan Ahmed Sial, Ramon Baldrich, Maria Vanrell,
- Abstract summary: Single image scene relighting aims to generate a realistic new version of an input image so that it appears to be illuminated by a new target light condition.
We propose two new datasets: a synthetic dataset with the ground truth of intrinsic components and a real dataset collected under laboratory conditions.
Our method outperforms the state-of-the-art methods in performance, as tested on both existing datasets and our newly developed datasets.
- Score: 0.7499722271664147
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Single image scene relighting aims to generate a realistic new version of an input image so that it appears to be illuminated by a new target light condition. Although existing works have explored this problem from various perspectives, generating relit images under arbitrary light conditions remains highly challenging, and related datasets are scarce. Our work addresses this problem from both the dataset and methodological perspectives. We propose two new datasets: a synthetic dataset with the ground truth of intrinsic components and a real dataset collected under laboratory conditions. These datasets alleviate the scarcity of existing datasets. To incorporate physical consistency in the relighting pipeline, we establish a two-stage network based on intrinsic decomposition, giving outputs at intermediate steps, thereby introducing physical constraints. When the training set lacks ground truth for intrinsic decomposition, we introduce an unsupervised module to ensure that the intrinsic outputs are satisfactory. Our method outperforms the state-of-the-art methods in performance, as tested on both existing datasets and our newly developed datasets. Furthermore, pretraining our method or other prior methods using our synthetic dataset can enhance their performance on other datasets. Since our method can accommodate any light conditions, it is capable of producing animated results. The dataset, method, and videos are publicly available.
Related papers
- Synthetic data enables faster annotation and robust segmentation for
multi-object grasping in clutter [9.092550803271005]
We propose a synthetic data generation method that minimizes human intervention and makes downstream image segmentation algorithms more robust.
Experiments show that the proposed synthetic scene generation can diminish labelling time dramatically.
Pick-and-place experiments demonstrate that segmentation trained on our hybrid dataset (98.9%, 70%) outperforms the real dataset and a publicly available dataset by (6.7%, 18.8%) and (2.8%, 10%) in terms of labelling and grasping success rate.
arXiv Detail & Related papers (2024-01-24T11:58:30Z) - Minimizing the Accumulated Trajectory Error to Improve Dataset
Distillation [151.70234052015948]
We propose a novel approach that encourages the optimization algorithm to seek a flat trajectory.
We show that the weights trained on synthetic data are robust against the accumulated errors perturbations with the regularization towards the flat trajectory.
Our method, called Flat Trajectory Distillation (FTD), is shown to boost the performance of gradient-matching methods by up to 4.7%.
arXiv Detail & Related papers (2022-11-20T15:49:11Z) - TRoVE: Transforming Road Scene Datasets into Photorealistic Virtual
Environments [84.6017003787244]
This work proposes a synthetic data generation pipeline to address the difficulties and domain-gaps present in simulated datasets.
We show that using annotations and visual cues from existing datasets, we can facilitate automated multi-modal data generation.
arXiv Detail & Related papers (2022-08-16T20:46:08Z) - DC-BENCH: Dataset Condensation Benchmark [79.18718490863908]
This work provides the first large-scale standardized benchmark on dataset condensation.
It consists of a suite of evaluations to comprehensively reflect the generability and effectiveness of condensation methods.
The benchmark library is open-sourced to facilitate future research and application.
arXiv Detail & Related papers (2022-07-20T03:54:05Z) - Delving into High-Quality Synthetic Face Occlusion Segmentation Datasets [83.749895930242]
We propose two techniques for producing high-quality naturalistic synthetic occluded faces.
We empirically show the effectiveness and robustness of both methods, even for unseen occlusions.
We present two high-resolution real-world occluded face datasets with fine-grained annotations, RealOcc and RealOcc-Wild.
arXiv Detail & Related papers (2022-05-12T17:03:57Z) - Partially fake it till you make it: mixing real and fake thermal images
for improved object detection [29.13557322147509]
We show the performance of the proposed system in the context of object detection in thermal videos.
Our single-modality detector achieves state-of-the-art results on the FLIR ADAS dataset.
arXiv Detail & Related papers (2021-06-25T12:56:09Z) - Joint Learning of Portrait Intrinsic Decomposition and Relighting [5.601217969637838]
Inverse rendering is the problem of decomposing an image into its intrinsic components, i.e. albedo, normal and lighting.
Here, we propose a new self-supervised training paradigm that reduces the need for full supervision on the decomposition task.
We show-case the effectiveness of our training paradigm on both intrinsic decomposition and relighting and demonstrate how the model struggles in both tasks without the self-supervised loss terms.
arXiv Detail & Related papers (2021-06-22T19:40:22Z) - Salient Objects in Clutter [130.63976772770368]
This paper identifies and addresses a serious design bias of existing salient object detection (SOD) datasets.
This design bias has led to a saturation in performance for state-of-the-art SOD models when evaluated on existing datasets.
We propose a new high-quality dataset and update the previous saliency benchmark.
arXiv Detail & Related papers (2021-05-07T03:49:26Z) - Optical Flow Dataset Synthesis from Unpaired Images [36.158607790844705]
We introduce a novel method to build a training set of pseudo-real images that can be used to train optical flow in a supervised manner.
Our dataset uses two unpaired frames from real data and creates pairs of frames by simulating random warps.
We thus obtain the benefit of directly training on real data while having access to an exact ground truth.
arXiv Detail & Related papers (2021-04-02T22:19:47Z) - OpenRooms: An End-to-End Open Framework for Photorealistic Indoor Scene
Datasets [103.54691385842314]
We propose a novel framework for creating large-scale photorealistic datasets of indoor scenes.
Our goal is to make the dataset creation process widely accessible.
This enables important applications in inverse rendering, scene understanding and robotics.
arXiv Detail & Related papers (2020-07-25T06:48:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.