Generative Modelling of BRDF Textures from Flash Images
- URL: http://arxiv.org/abs/2102.11861v1
- Date: Tue, 23 Feb 2021 18:45:18 GMT
- Title: Generative Modelling of BRDF Textures from Flash Images
- Authors: Philipp Henzler, Valentin Deschaintre, Niloy J. Mitra, Tobias Ritschel
- Abstract summary: We learn a latent space for easy capture, semantic editing, consistent, and efficient reproduction of visual material appearance.
In a second step, conditioned on the material code, our method produces an infinite and diverse spatial field of BRDF model parameters.
- Score: 50.660026124025265
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We learn a latent space for easy capture, semantic editing, consistent
interpolation, and efficient reproduction of visual material appearance. When
users provide a photo of a stationary natural material captured under flash
light illumination, it is converted in milliseconds into a latent material
code. In a second step, conditioned on the material code, our method, again in
milliseconds, produces an infinite and diverse spatial field of BRDF model
parameters (diffuse albedo, specular albedo, roughness, normals) that allows
rendering in complex scenes and illuminations, matching the appearance of the
input picture. Technically, we jointly embed all flash images into a latent
space using a convolutional encoder, and -- conditioned on these latent codes
-- convert random spatial fields into fields of BRDF parameters using a
convolutional neural network (CNN). We condition these BRDF parameters to match
the visual characteristics (statistics and spectra of visual features) of the
input under matching light. A user study confirms that the semantics of the
latent material space agree with user expectations and compares our approach
favorably to previous work.
Related papers
- Neural Appearance Modeling From Single Images [3.3090362820994526]
We propose a material appearance modeling neural network for visualizing plausible, spatially-varying materials under diverse view and lighting conditions.
Our network is composed of two network stages: a network that infers learned per-pixel neural parameters of a material from a single input photograph, and a network that renders the material.
arXiv Detail & Related papers (2024-06-08T18:56:03Z) - Unified Multi-Modal Latent Diffusion for Joint Subject and Text
Conditional Image Generation [63.061871048769596]
We present a novel Unified Multi-Modal Latent Diffusion (UMM-Diffusion) which takes joint texts and images containing specified subjects as input sequences.
To be more specific, both input texts and images are encoded into one unified multi-modal latent space.
Our method is able to generate high-quality images with complex semantics from both aspects of input texts and images.
arXiv Detail & Related papers (2023-03-16T13:50:20Z) - Light Field Implicit Representation for Flexible Resolution
Reconstruction [9.173467982128514]
We propose an implicit representation model for 4D light fields conditioned on a sparse set of input views.
Our model is trained to output the light field values for a continuous range of coordinates.
Experiments show that our method achieves state-of-the-art performance for the synthesis of view while being computationally fast.
arXiv Detail & Related papers (2021-11-30T23:59:02Z) - Cycle-Consistent Inverse GAN for Text-to-Image Synthesis [101.97397967958722]
We propose a novel unified framework of Cycle-consistent Inverse GAN for both text-to-image generation and text-guided image manipulation tasks.
We learn a GAN inversion model to convert the images back to the GAN latent space and obtain the inverted latent codes for each image.
In the text-guided optimization module, we generate images with the desired semantic attributes by optimizing the inverted latent codes.
arXiv Detail & Related papers (2021-08-03T08:38:16Z) - Learning optical flow from still images [53.295332513139925]
We introduce a framework to generate accurate ground-truth optical flow annotations quickly and in large amounts from any readily available single real picture.
We virtually move the camera in the reconstructed environment with known motion vectors and rotation angles.
When trained with our data, state-of-the-art optical flow networks achieve superior generalization to unseen real data.
arXiv Detail & Related papers (2021-04-08T17:59:58Z) - PhySG: Inverse Rendering with Spherical Gaussians for Physics-based
Material Editing and Relighting [60.75436852495868]
We present PhySG, an inverse rendering pipeline that reconstructs geometry, materials, and illumination from scratch from RGB input images.
We demonstrate, with both synthetic and real data, that our reconstructions not only enable rendering of novel viewpoints, but also physics-based appearance editing of materials and illumination.
arXiv Detail & Related papers (2021-04-01T17:59:02Z) - Convolutional Autoencoder for Blind Hyperspectral Image Unmixing [0.0]
spectral unmixing is a technique to decompose a mixed pixel into two fundamental representatives: endmembers and abundances.
In this paper, a novel architecture is proposed to perform blind unmixing on hyperspectral images.
arXiv Detail & Related papers (2020-11-18T17:41:31Z) - Two-shot Spatially-varying BRDF and Shape Estimation [89.29020624201708]
We propose a novel deep learning architecture with a stage-wise estimation of shape and SVBRDF.
We create a large-scale synthetic training dataset with domain-randomized geometry and realistic materials.
Experiments on both synthetic and real-world datasets show that our network trained on a synthetic dataset can generalize well to real-world images.
arXiv Detail & Related papers (2020-04-01T12:56:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.