One-shot recognition of any material anywhere using contrastive learning
with physics-based rendering
- URL: http://arxiv.org/abs/2212.00648v1
- Date: Thu, 1 Dec 2022 16:49:53 GMT
- Title: One-shot recognition of any material anywhere using contrastive learning
with physics-based rendering
- Authors: Manuel S. Drehwald (3), Sagi Eppel (1 and 2 and 4), Jolina Li (2 and
4), Han Hao (2), Alan Aspuru-Guzik (1 and 2) ((1) Vector institute, (2)
University of Toronto, (3) Karlsruhe Institute of Technology, (4) Innoviz)
- Abstract summary: We present MatSim: a synthetic dataset, a benchmark, and a method for computer vision based recognition of similarities and transitions between materials and textures.
The visual recognition of materials is essential to everything from examining food while cooking to inspecting agriculture, chemistry, and industrial products.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We present MatSim: a synthetic dataset, a benchmark, and a method for
computer vision based recognition of similarities and transitions between
materials and textures, focusing on identifying any material under any
conditions using one or a few examples (one-shot learning). The visual
recognition of materials is essential to everything from examining food while
cooking to inspecting agriculture, chemistry, and industrial products. In this
work, we utilize giant repositories used by computer graphics artists to
generate a new CGI dataset for material similarity. We use physics-based
rendering (PBR) repositories for visual material simulation, assign these
materials random 3D objects, and render images with a vast range of backgrounds
and illumination conditions (HDRI). We add a gradual transition between
materials to support applications with a smooth transition between states (like
gradually cooked food). We also render materials inside transparent containers
to support beverage and chemistry lab use cases. We then train a contrastive
learning network to generate a descriptor that identifies unfamiliar materials
using a single image. We also present a new benchmark for a few-shot material
recognition that contains a wide range of real-world examples, including the
state of a chemical reaction, rotten/fresh fruits, states of food, different
types of construction materials, types of ground, and many other use cases
involving material states, transitions and subclasses. We show that a network
trained on the MatSim synthetic dataset outperforms state-of-the-art models
like Clip on the benchmark, despite being tested on material classes that were
not seen during training. The dataset, benchmark, code and trained models are
available online.
Related papers
- MaPa: Text-driven Photorealistic Material Painting for 3D Shapes [80.66880375862628]
This paper aims to generate materials for 3D meshes from text descriptions.
Unlike existing methods that synthesize texture maps, we propose to generate segment-wise procedural material graphs.
Our framework supports high-quality rendering and provides substantial flexibility in editing.
arXiv Detail & Related papers (2024-04-26T17:54:38Z) - Make-it-Real: Unleashing Large Multimodal Model for Painting 3D Objects with Realistic Materials [108.59709545364395]
GPT-4V can effectively recognize and describe materials, allowing the construction of a detailed material library.
The correctly matched materials are then meticulously applied as reference for the new SVBRDF material generation.
Make-it-Real offers a streamlined integration into the 3D content creation workflow.
arXiv Detail & Related papers (2024-04-25T17:59:58Z) - MaterialSeg3D: Segmenting Dense Materials from 2D Priors for 3D Assets [63.284244910964475]
We propose a 3D asset material generation framework to infer underlying material from the 2D semantic prior.
Based on such a prior model, we devise a mechanism to parse material in 3D space.
arXiv Detail & Related papers (2024-04-22T07:00:17Z) - Intrinsic Image Diffusion for Indoor Single-view Material Estimation [55.276815106443976]
We present Intrinsic Image Diffusion, a generative model for appearance decomposition of indoor scenes.
Given a single input view, we sample multiple possible material explanations represented as albedo, roughness, and metallic maps.
Our method produces significantly sharper, more consistent, and more detailed materials, outperforming state-of-the-art methods by $1.5dB$ on PSNR and by $45%$ better FID score on albedo prediction.
arXiv Detail & Related papers (2023-12-19T15:56:19Z) - Alchemist: Parametric Control of Material Properties with Diffusion
Models [51.63031820280475]
Our method capitalizes on the generative prior of text-to-image models known for photorealism.
We show the potential application of our model to material edited NeRFs.
arXiv Detail & Related papers (2023-12-05T18:58:26Z) - Material Palette: Extraction of Materials from a Single Image [19.410479434979493]
We propose a method to extract physically-based rendering (PBR) materials from a single real-world image.
We map regions of the image to material concepts using a diffusion model, which allows the sampling of texture images resembling each material in the scene.
Second, we benefit from a separate network to decompose the generated textures into Spatially Varying BRDFs.
arXiv Detail & Related papers (2023-11-28T18:59:58Z) - MatFuse: Controllable Material Generation with Diffusion Models [10.993516790237503]
MatFuse is a unified approach that harnesses the generative power of diffusion models for creation and editing of 3D materials.
Our method integrates multiple sources of conditioning, including color palettes, sketches, text, and pictures, enhancing creative possibilities.
We demonstrate the effectiveness of MatFuse under multiple conditioning settings and explore the potential of material editing.
arXiv Detail & Related papers (2023-08-22T12:54:48Z) - PhotoMat: A Material Generator Learned from Single Flash Photos [37.42765147463852]
Previous generative models for materials have been trained exclusively on synthetic data.
We propose PhotoMat: the first material generator trained exclusively on real photos of material samples captured using a cell phone camera with flash.
We show that our generated materials have better visual quality than previous material generators trained on synthetic data.
arXiv Detail & Related papers (2023-05-20T22:27:41Z) - Neural Photometry-guided Visual Attribute Transfer [4.630419389180576]
We present a deep learning-based method for propagating visual material attributes to larger samples of the same or similar materials.
For training, we leverage images of the material taken under multiple illuminations and a dedicated data augmentation policy.
Our model relies on a supervised image-to-image translation framework and is agnostic to the transferred domain.
arXiv Detail & Related papers (2021-12-05T09:22:28Z) - The joint role of geometry and illumination on material recognition [16.01513204879645]
We study how the interplay of geometry, illumination, and their spatial frequencies affects human performance on material recognition tasks.
We train a deep neural network on material recognition tasks to accurately classify materials.
arXiv Detail & Related papers (2021-01-07T11:29:52Z) - Fed-Sim: Federated Simulation for Medical Imaging [131.56325440976207]
We introduce a physics-driven generative approach that consists of two learnable neural modules.
We show that our data synthesis framework improves the downstream segmentation performance on several datasets.
arXiv Detail & Related papers (2020-09-01T19:17:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.