Differential Viewpoints for Ground Terrain Material Recognition
- URL: http://arxiv.org/abs/2009.11072v1
- Date: Tue, 22 Sep 2020 02:57:28 GMT
- Title: Differential Viewpoints for Ground Terrain Material Recognition
- Authors: Jia Xue, Hang Zhang, Ko Nishino, Kristin J. Dana
- Abstract summary: We build a large-scale material database to support ground terrain recognition for applications such as autonomous driving and robot navigation.
We develop a novel approach for material recognition called texture-encoded angular network (TEAN) that combines deep encoding of RGB information and differential angular images for angular-gradient features.
Our results show that TEAN achieves recognition performance that surpasses single view performance and standard (non-differential/large-angle sampling) multiview performance.
- Score: 32.91058153755717
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Computational surface modeling that underlies material recognition has
transitioned from reflectance modeling using in-lab controlled radiometric
measurements to image-based representations based on internet-mined single-view
images captured in the scene. We take a middle-ground approach for material
recognition that takes advantage of both rich radiometric cues and flexible
image capture. A key concept is differential angular imaging, where small
angular variations in image capture enables angular-gradient features for an
enhanced appearance representation that improves recognition. We build a
large-scale material database, Ground Terrain in Outdoor Scenes (GTOS)
database, to support ground terrain recognition for applications such as
autonomous driving and robot navigation. The database consists of over 30,000
images covering 40 classes of outdoor ground terrain under varying weather and
lighting conditions. We develop a novel approach for material recognition
called texture-encoded angular network (TEAN) that combines deep encoding
pooling of RGB information and differential angular images for angular-gradient
features to fully leverage this large dataset. With this novel network
architecture, we extract characteristics of materials encoded in the angular
and spatial gradients of their appearance. Our results show that TEAN achieves
recognition performance that surpasses single view performance and standard
(non-differential/large-angle sampling) multiview performance.
Related papers
- RMAFF-PSN: A Residual Multi-Scale Attention Feature Fusion Photometric Stereo Network [37.759675702107586]
Predicting accurate maps of objects from two-dimensional images in regions of complex structure spatial material variations is challenging.
We propose a method of calibrated feature information from different resolution stages and scales of the image.
This approach preserves more physical information, such as texture and geometry of the object in complex regions.
arXiv Detail & Related papers (2024-04-11T14:05:37Z) - Multi-Projection Fusion and Refinement Network for Salient Object
Detection in 360{\deg} Omnidirectional Image [141.10227079090419]
We propose a Multi-Projection Fusion and Refinement Network (MPFR-Net) to detect the salient objects in 360deg omnidirectional image.
MPFR-Net uses the equirectangular projection image and four corresponding cube-unfolding images as inputs.
Experimental results on two omnidirectional datasets demonstrate that the proposed approach outperforms the state-of-the-art methods both qualitatively and quantitatively.
arXiv Detail & Related papers (2022-12-23T14:50:40Z) - Joint Learning of Deep Texture and High-Frequency Features for
Computer-Generated Image Detection [24.098604827919203]
We propose a joint learning strategy with deep texture and high-frequency features for CG image detection.
A semantic segmentation map is generated to guide the affine transformation operation.
The combination of the original image and the high-frequency components of the original and rendered images are fed into a multi-branch neural network equipped with attention mechanisms.
arXiv Detail & Related papers (2022-09-07T17:30:40Z) - Extracting Triangular 3D Models, Materials, and Lighting From Images [59.33666140713829]
We present an efficient method for joint optimization of materials and lighting from multi-view image observations.
We leverage meshes with spatially-varying materials and environment that can be deployed in any traditional graphics engine.
arXiv Detail & Related papers (2021-11-24T13:58:20Z) - Neural Radiance Fields Approach to Deep Multi-View Photometric Stereo [103.08512487830669]
We present a modern solution to the multi-view photometric stereo problem (MVPS)
We procure the surface orientation using a photometric stereo (PS) image formation model and blend it with a multi-view neural radiance field representation to recover the object's surface geometry.
Our method performs neural rendering of multi-view images while utilizing surface normals estimated by a deep photometric stereo network.
arXiv Detail & Related papers (2021-10-11T20:20:03Z) - Light Field Reconstruction Using Convolutional Network on EPI and
Extended Applications [78.63280020581662]
A novel convolutional neural network (CNN)-based framework is developed for light field reconstruction from a sparse set of views.
We demonstrate the high performance and robustness of the proposed framework compared with state-of-the-art algorithms.
arXiv Detail & Related papers (2021-03-24T08:16:32Z) - VAE-Info-cGAN: Generating Synthetic Images by Combining Pixel-level and
Feature-level Geospatial Conditional Inputs [0.0]
We present a conditional generative model for synthesizing semantically rich images simultaneously conditioned on a pixellevel (PLC) and a featurelevel condition (FLC)
Experiments on a GPS dataset show that the proposed model can accurately generate various forms of macroscopic aggregates across different geographic locations.
arXiv Detail & Related papers (2020-12-08T03:46:19Z) - Angular Luminance for Material Segmentation [6.374538197161135]
Moving cameras provide multiple intensity measurements per pixel, yet often semantic segmentation, material recognition, and object recognition do not utilize this information.
We utilize per-pixel angular luminance distributions as a key feature in discriminating the material of the surface.
For real-world materials there is significant intra-class variation that can be managed by building a angular luminance network (AngLNet)
arXiv Detail & Related papers (2020-09-22T21:15:27Z) - City-Scale Visual Place Recognition with Deep Local Features Based on
Multi-Scale Ordered VLAD Pooling [5.274399407597545]
We present a fully-automated system for place recognition at a city-scale based on content-based image retrieval.
Firstly, we take a comprehensive analysis of visual place recognition and sketch out the unique challenges of the task.
Next, we propose yet a simple pooling approach on top of convolutional neural network activations to embed the spatial information into the image representation vector.
arXiv Detail & Related papers (2020-09-19T15:21:59Z) - Two-shot Spatially-varying BRDF and Shape Estimation [89.29020624201708]
We propose a novel deep learning architecture with a stage-wise estimation of shape and SVBRDF.
We create a large-scale synthetic training dataset with domain-randomized geometry and realistic materials.
Experiments on both synthetic and real-world datasets show that our network trained on a synthetic dataset can generalize well to real-world images.
arXiv Detail & Related papers (2020-04-01T12:56:13Z) - Deep 3D Capture: Geometry and Reflectance from Sparse Multi-View Images [59.906948203578544]
We introduce a novel learning-based method to reconstruct the high-quality geometry and complex, spatially-varying BRDF of an arbitrary object.
We first estimate per-view depth maps using a deep multi-view stereo network.
These depth maps are used to coarsely align the different views.
We propose a novel multi-view reflectance estimation network architecture.
arXiv Detail & Related papers (2020-03-27T21:28:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.