Multi-View Photometric Stereo: A Robust Solution and Benchmark Dataset
for Spatially Varying Isotropic Materials
- URL: http://arxiv.org/abs/2001.06659v1
- Date: Sat, 18 Jan 2020 12:26:22 GMT
- Title: Multi-View Photometric Stereo: A Robust Solution and Benchmark Dataset
for Spatially Varying Isotropic Materials
- Authors: Min Li, Zhenglong Zhou, Zhe Wu, Boxin Shi, Changyu Diao, and Ping Tan
- Abstract summary: We present a method to capture both 3D shape and spatially varying reflectance with a multi-view photometric stereo technique.
Our algorithm is suitable for perspective cameras and nearby point light sources.
- Score: 65.95928593628128
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present a method to capture both 3D shape and spatially varying
reflectance with a multi-view photometric stereo (MVPS) technique that works
for general isotropic materials. Our algorithm is suitable for perspective
cameras and nearby point light sources. Our data capture setup is simple, which
consists of only a digital camera, some LED lights, and an optional automatic
turntable. From a single viewpoint, we use a set of photometric stereo images
to identify surface points with the same distance to the camera. We collect
this information from multiple viewpoints and combine it with
structure-from-motion to obtain a precise reconstruction of the complete 3D
shape. The spatially varying isotropic bidirectional reflectance distribution
function (BRDF) is captured by simultaneously inferring a set of basis BRDFs
and their mixing weights at each surface point. In experiments, we demonstrate
our algorithm with two different setups: a studio setup for highest precision
and a desktop setup for best usability. According to our experiments, under the
studio setting, the captured shapes are accurate to 0.5 millimeters and the
captured reflectance has a relative root-mean-square error (RMSE) of 9%. We
also quantitatively evaluate state-of-the-art MVPS on a newly collected
benchmark dataset, which is publicly available for inspiring future research.
Related papers
- VFMM3D: Releasing the Potential of Image by Vision Foundation Model for Monocular 3D Object Detection [80.62052650370416]
monocular 3D object detection holds significant importance across various applications, including autonomous driving and robotics.
In this paper, we present VFMM3D, an innovative framework that leverages the capabilities of Vision Foundation Models (VFMs) to accurately transform single-view images into LiDAR point cloud representations.
arXiv Detail & Related papers (2024-04-15T03:12:12Z) - Multi-Modal Dataset Acquisition for Photometrically Challenging Object [56.30027922063559]
This paper addresses the limitations of current datasets for 3D vision tasks in terms of accuracy, size, realism, and suitable imaging modalities for photometrically challenging objects.
We propose a novel annotation and acquisition pipeline that enhances existing 3D perception and 6D object pose datasets.
arXiv Detail & Related papers (2023-08-21T10:38:32Z) - A CNN Based Approach for the Point-Light Photometric Stereo Problem [26.958763133729846]
We propose a CNN-based approach capable of handling realistic assumptions by leveraging recent improvements of deep neural networks for far-field Photometric Stereo.
Our approach outperforms the state-of-the-art on the DiLiGenT real world dataset.
In order to measure the performance of our approach for near-field point-light source PS data, we introduce LUCES the first real-world 'dataset for near-fieLd point light soUrCe photomEtric Stereo'
arXiv Detail & Related papers (2022-10-10T12:57:12Z) - Sparse Ellipsometry: Portable Acquisition of Polarimetric SVBRDF and
Shape with Unstructured Flash Photography [32.68190169944569]
We present a portable polarimetric acquisition method that captures both polarimetric SVBRDF and 3D shape simultaneously.
Instead of days, the total acquisition time varies between twenty and thirty minutes per object.
Our results show a strong agreement with a recent ground-truth dataset of captured polarimetric BRDFs of real-world objects.
arXiv Detail & Related papers (2022-07-09T09:42:59Z) - Neural Radiance Fields Approach to Deep Multi-View Photometric Stereo [103.08512487830669]
We present a modern solution to the multi-view photometric stereo problem (MVPS)
We procure the surface orientation using a photometric stereo (PS) image formation model and blend it with a multi-view neural radiance field representation to recover the object's surface geometry.
Our method performs neural rendering of multi-view images while utilizing surface normals estimated by a deep photometric stereo network.
arXiv Detail & Related papers (2021-10-11T20:20:03Z) - LUCES: A Dataset for Near-Field Point Light Source Photometric Stereo [30.31403197697561]
We introduce LUCES, the first real-world 'dataset for near-fieLd point light soUrCe photomEtric Stereo' of 14 objects of a varying of materials.
A device counting 52 LEDs has been designed to lit each object positioned 10 to 30 centimeters away from the camera.
We evaluate the performance of the latest near-field Photometric Stereo algorithms on the proposed dataset.
arXiv Detail & Related papers (2021-04-27T12:30:42Z) - Leveraging Spatial and Photometric Context for Calibrated Non-Lambertian
Photometric Stereo [61.6260594326246]
We introduce an efficient fully-convolutional architecture that can leverage both spatial and photometric context simultaneously.
Using separable 4D convolutions and 2D heat-maps reduces the size and makes more efficient.
arXiv Detail & Related papers (2021-03-22T18:06:58Z) - Event-based Stereo Visual Odometry [42.77238738150496]
We present a solution to the problem of visual odometry from the data acquired by a stereo event-based camera rig.
We seek to maximize thetemporal consistency of stereo event-based data while using a simple and efficient representation.
arXiv Detail & Related papers (2020-07-30T15:53:28Z) - Lightweight Multi-View 3D Pose Estimation through Camera-Disentangled
Representation [57.11299763566534]
We present a solution to recover 3D pose from multi-view images captured with spatially calibrated cameras.
We exploit 3D geometry to fuse input images into a unified latent representation of pose, which is disentangled from camera view-points.
Our architecture then conditions the learned representation on camera projection operators to produce accurate per-view 2d detections.
arXiv Detail & Related papers (2020-04-05T12:52:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.