Scene Inference for Object Illumination Editing
- URL: http://arxiv.org/abs/2108.00150v1
- Date: Sat, 31 Jul 2021 05:02:52 GMT
- Title: Scene Inference for Object Illumination Editing
- Authors: Zhongyun Bao, Chengjiang Long, Gang Fu, Daquan Liu, Yuanzhen Li,
Jiaming Wu, Chunxia Xiao
- Abstract summary: We apply a physically-based rendering method to create a large-scale, high-quality dataset, named IH dataset.
We also propose a deep learning-based SI-GAN method, a multi-task collaborative network, to edit object illumination.
Our proposed SI-GAN provides a practical and effective solution for image-based object illumination editing, and validate the superiority of our method against state-of-the-art methods.
- Score: 24.529871334658573
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The seamless illumination integration between a foreground object and a
background scene is an important but challenging task in computer vision and
augmented reality community. However, to our knowledge, there is no publicly
available high-quality dataset that meets the illumination seamless integration
task, which greatly hinders the development of this research direction. To this
end, we apply a physically-based rendering method to create a large-scale,
high-quality dataset, named IH dataset, which provides rich illumination
information for seamless illumination integration task. In addition, we propose
a deep learning-based SI-GAN method, a multi-task collaborative network, which
makes full use of the multi-scale attention mechanism and adversarial learning
strategy to directly infer mapping relationship between the inserted foreground
object and corresponding background environment, and edit object illumination
according to the proposed illumination exchange mechanism in parallel network.
By this means, we can achieve the seamless illumination integration without
explicit estimation of 3D geometric information. Comprehensive experiments on
both our dataset and real-world images collected from the Internet show that
our proposed SI-GAN provides a practical and effective solution for image-based
object illumination editing, and validate the superiority of our method against
state-of-the-art methods.
Related papers
- HUE Dataset: High-Resolution Event and Frame Sequences for Low-Light Vision [16.432164340779266]
We introduce the HUE dataset, a collection of high-resolution event and frame sequences captured in low-light conditions.
Our dataset includes 106 sequences, encompassing indoor, cityscape, twilight, night, driving, and controlled scenarios.
We employ both qualitative and quantitative evaluations to assess state-of-the-art low-light enhancement and event-based image reconstruction methods.
arXiv Detail & Related papers (2024-10-24T21:15:15Z) - GS-Phong: Meta-Learned 3D Gaussians for Relightable Novel View Synthesis [63.5925701087252]
We propose a novel method for representing a scene illuminated by a point light using a set of relightable 3D Gaussian points.
Inspired by the Blinn-Phong model, our approach decomposes the scene into ambient, diffuse, and specular components.
To facilitate the decomposition of geometric information independent of lighting conditions, we introduce a novel bilevel optimization-based meta-learning framework.
arXiv Detail & Related papers (2024-05-31T13:48:54Z) - Multi-Modal Dataset Acquisition for Photometrically Challenging Object [56.30027922063559]
This paper addresses the limitations of current datasets for 3D vision tasks in terms of accuracy, size, realism, and suitable imaging modalities for photometrically challenging objects.
We propose a novel annotation and acquisition pipeline that enhances existing 3D perception and 6D object pose datasets.
arXiv Detail & Related papers (2023-08-21T10:38:32Z) - Multitask AET with Orthogonal Tangent Regularity for Dark Object
Detection [84.52197307286681]
We propose a novel multitask auto encoding transformation (MAET) model to enhance object detection in a dark environment.
In a self-supervision manner, the MAET learns the intrinsic visual structure by encoding and decoding the realistic illumination-degrading transformation.
We have achieved the state-of-the-art performance using synthetic and real-world datasets.
arXiv Detail & Related papers (2022-05-06T16:27:14Z) - Salient Objects in Clutter [130.63976772770368]
This paper identifies and addresses a serious design bias of existing salient object detection (SOD) datasets.
This design bias has led to a saturation in performance for state-of-the-art SOD models when evaluated on existing datasets.
We propose a new high-quality dataset and update the previous saliency benchmark.
arXiv Detail & Related papers (2021-05-07T03:49:26Z) - Lighting the Darkness in the Deep Learning Era [118.35081853500411]
Low-light image enhancement (LLIE) aims at improving the perception or interpretability of an image captured in an environment with poor illumination.
Recent advances in this area are dominated by deep learning-based solutions.
We provide a comprehensive survey to cover various aspects ranging from algorithm taxonomy to unsolved open issues.
arXiv Detail & Related papers (2021-04-21T19:12:19Z) - Dynamic Fusion Network For Light Field Depth Estimation [32.64928379844675]
We propose a dynamically multi modal learning strategy which incorporates RGB data and the focal stack in our framework.
The success of our method is demonstrated by achieving the state of the art performance on two datasets.
arXiv Detail & Related papers (2021-04-13T06:45:11Z) - PX-NET: Simple and Efficient Pixel-Wise Training of Photometric Stereo
Networks [26.958763133729846]
Retrieving accurate 3D reconstructions of objects from the way they reflect light is a very challenging task in computer vision.
We propose a novel pixel-wise training procedure for normal prediction by replacing the training data (observation maps) of globally rendered images with independent per-pixel generated data.
Our network, PX-NET, achieves the state-of-the-art performance compared to other pixelwise methods on synthetic datasets.
arXiv Detail & Related papers (2020-08-11T18:03:13Z) - OpenRooms: An End-to-End Open Framework for Photorealistic Indoor Scene
Datasets [103.54691385842314]
We propose a novel framework for creating large-scale photorealistic datasets of indoor scenes.
Our goal is to make the dataset creation process widely accessible.
This enables important applications in inverse rendering, scene understanding and robotics.
arXiv Detail & Related papers (2020-07-25T06:48:47Z) - IllumiNet: Transferring Illumination from Planar Surfaces to Virtual
Objects in Augmented Reality [38.83696624634213]
This paper presents an illumination estimation method for virtual objects in real environment by learning.
Given a single RGB image, our method directly infers the relit virtual object by transferring the illumination features extracted from planar surfaces in the scene to the desired geometries.
arXiv Detail & Related papers (2020-07-12T13:11:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.