Two-shot Spatially-varying BRDF and Shape Estimation
- URL: http://arxiv.org/abs/2004.00403v1
- Date: Wed, 1 Apr 2020 12:56:13 GMT
- Title: Two-shot Spatially-varying BRDF and Shape Estimation
- Authors: Mark Boss, Varun Jampani, Kihwan Kim, Hendrik P.A. Lensch, Jan Kautz
- Abstract summary: We propose a novel deep learning architecture with a stage-wise estimation of shape and SVBRDF.
We create a large-scale synthetic training dataset with domain-randomized geometry and realistic materials.
Experiments on both synthetic and real-world datasets show that our network trained on a synthetic dataset can generalize well to real-world images.
- Score: 89.29020624201708
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Capturing the shape and spatially-varying appearance (SVBRDF) of an object
from images is a challenging task that has applications in both computer vision
and graphics. Traditional optimization-based approaches often need a large
number of images taken from multiple views in a controlled environment. Newer
deep learning-based approaches require only a few input images, but the
reconstruction quality is not on par with optimization techniques. We propose a
novel deep learning architecture with a stage-wise estimation of shape and
SVBRDF. The previous predictions guide each estimation, and a joint refinement
network later refines both SVBRDF and shape. We follow a practical mobile image
capture setting and use unaligned two-shot flash and no-flash images as input.
Both our two-shot image capture and network inference can run on mobile
hardware. We also create a large-scale synthetic training dataset with
domain-randomized geometry and realistic materials. Extensive experiments on
both synthetic and real-world datasets show that our network trained on a
synthetic dataset can generalize well to real-world images. Comparisons with
recent approaches demonstrate the superior performance of the proposed
approach.
Related papers
- Deep Domain Adaptation: A Sim2Real Neural Approach for Improving Eye-Tracking Systems [80.62854148838359]
Eye image segmentation is a critical step in eye tracking that has great influence over the final gaze estimate.
We use dimensionality-reduction techniques to measure the overlap between the target eye images and synthetic training data.
Our methods result in robust, improved performance when tackling the discrepancy between simulation and real-world data samples.
arXiv Detail & Related papers (2024-03-23T22:32:06Z) - Spectral Graphormer: Spectral Graph-based Transformer for Egocentric
Two-Hand Reconstruction using Multi-View Color Images [33.70056950818641]
We propose a novel transformer-based framework that reconstructs two high fidelity hands from multi-view RGB images.
We show that our framework is able to produce realistic two-hand reconstructions and demonstrate the generalisation of synthetic-trained models to real data.
arXiv Detail & Related papers (2023-08-21T20:07:02Z) - Improving Human-Object Interaction Detection via Virtual Image Learning [68.56682347374422]
Human-Object Interaction (HOI) detection aims to understand the interactions between humans and objects.
In this paper, we propose to alleviate the impact of such an unbalanced distribution via Virtual Image Leaning (VIL)
A novel label-to-image approach, Multiple Steps Image Creation (MUSIC), is proposed to create a high-quality dataset that has a consistent distribution with real images.
arXiv Detail & Related papers (2023-08-04T10:28:48Z) - Rethinking Motion Deblurring Training: A Segmentation-Based Method for
Simulating Non-Uniform Motion Blurred Images [0.0]
We propose an efficient procedural methodology to generate sharp/blurred image pairs.
This allows generating virtually unlimited realistic and diverse training pairs.
We observed superior generalization performance for the ultimate task of deblurring real motion-blurred photos.
arXiv Detail & Related papers (2022-09-26T13:20:35Z) - Learning Deformable Image Registration from Optimization: Perspective,
Modules, Bilevel Training and Beyond [62.730497582218284]
We develop a new deep learning based framework to optimize a diffeomorphic model via multi-scale propagation.
We conduct two groups of image registration experiments on 3D volume datasets including image-to-atlas registration on brain MRI data and image-to-image registration on liver CT data.
arXiv Detail & Related papers (2020-04-30T03:23:45Z) - Deep CG2Real: Synthetic-to-Real Translation via Image Disentanglement [78.58603635621591]
Training an unpaired synthetic-to-real translation network in image space is severely under-constrained.
We propose a semi-supervised approach that operates on the disentangled shading and albedo layers of the image.
Our two-stage pipeline first learns to predict accurate shading in a supervised fashion using physically-based renderings as targets.
arXiv Detail & Related papers (2020-03-27T21:45:41Z) - Contextual Encoder-Decoder Network for Visual Saliency Prediction [42.047816176307066]
We propose an approach based on a convolutional neural network pre-trained on a large-scale image classification task.
We combine the resulting representations with global scene information for accurately predicting visual saliency.
Compared to state of the art approaches, the network is based on a lightweight image classification backbone.
arXiv Detail & Related papers (2019-02-18T16:15:25Z) - Scene Text Synthesis for Efficient and Effective Deep Network Training [62.631176120557136]
We develop an innovative image synthesis technique that composes annotated training images by embedding foreground objects of interest into background images.
The proposed technique consists of two key components that in principle boost the usefulness of the synthesized images in deep network training.
Experiments over a number of public datasets demonstrate the effectiveness of our proposed image synthesis technique.
arXiv Detail & Related papers (2019-01-26T10:15:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.