Few-shot Non-line-of-sight Imaging with Signal-surface Collaborative
Regularization
- URL: http://arxiv.org/abs/2211.15367v1
- Date: Mon, 21 Nov 2022 11:19:20 GMT
- Title: Few-shot Non-line-of-sight Imaging with Signal-surface Collaborative
Regularization
- Authors: Xintong Liu, Jianyu Wang, Leping Xiao, Xing Fu, Lingyun Qiu, Zuoqiang
Shi
- Abstract summary: Non-line-of-sight imaging technique aims to reconstruct targets from multiply reflected light.
We propose a signal-surface collaborative regularization framework that provides noise-robust reconstructions with a minimal number of measurements.
Our approach has great potential in real-time non-line-of-sight imaging applications such as rescue operations and autonomous driving.
- Score: 18.466941045530408
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The non-line-of-sight imaging technique aims to reconstruct targets from
multiply reflected light. For most existing methods, dense points on the relay
surface are raster scanned to obtain high-quality reconstructions, which
requires a long acquisition time. In this work, we propose a signal-surface
collaborative regularization (SSCR) framework that provides noise-robust
reconstructions with a minimal number of measurements. Using Bayesian
inference, we design joint regularizations of the estimated signal, the 3D
voxel-based representation of the objects, and the 2D surface-based description
of the targets. To our best knowledge, this is the first work that combines
regularizations in mixed dimensions for hidden targets. Experiments on
synthetic and experimental datasets illustrated the efficiency and robustness
of the proposed method under both confocal and non-confocal settings. We report
the reconstruction of the hidden targets with complex geometric structures with
only $5 \times 5$ confocal measurements from public datasets, indicating an
acceleration of the conventional measurement process by a factor of 10000.
Besides, the proposed method enjoys low time and memory complexities with
sparse measurements. Our approach has great potential in real-time
non-line-of-sight imaging applications such as rescue operations and autonomous
driving.
Related papers
- PF3plat: Pose-Free Feed-Forward 3D Gaussian Splatting [54.7468067660037]
PF3plat sets a new state-of-the-art across all benchmarks, supported by comprehensive ablation studies validating our design choices.
Our framework capitalizes on fast speed, scalability, and high-quality 3D reconstruction and view synthesis capabilities of 3DGS.
arXiv Detail & Related papers (2024-10-29T15:28:15Z) - Distributed Stochastic Optimization of a Neural Representation Network for Time-Space Tomography Reconstruction [4.689071714940848]
4D time-space reconstruction of dynamic events or deforming objects using Xray computed tomography (CT) is an extremely ill-posed inverse problem.
Existing approaches assume that the object remains static for the duration of several tens or hundreds of X-ray projection measurement images.
We propose to perform a 4D time-space reconstruction using a distributed implicit neural representation network that is trained using a novel distributed training algorithm.
arXiv Detail & Related papers (2024-04-29T19:41:51Z) - Stable Surface Regularization for Fast Few-Shot NeRF [76.00444039563581]
We develop a stable surface regularization technique called Annealing Signed Distance Function (ASDF)
We observe that the Eikonal loss requires dense training signal to shape different level-sets of SDF, leading to low-fidelity results under few-shot training.
The proposed approach is up to 45 times faster than existing few-shot novel view synthesis methods.
arXiv Detail & Related papers (2024-03-29T05:39:47Z) - Cryo-forum: A framework for orientation recovery with uncertainty
measure with the application in cryo-EM image analysis [0.0]
This paper introduces a novel approach that uses a 10-dimensional feature vector to represent the orientation and applies a Quadratically-Constrained Quadratic Program to derive the predicted orientation as a unit quaternion, supplemented by an uncertainty metric.
Our numerical analysis demonstrates that our methodology effectively recovers orientations from 2D cryo-EM images in an end-to-end manner. Importantly, the inclusion of uncertainty allows for direct clean-up of the dataset at the 3D level.
arXiv Detail & Related papers (2023-07-19T09:09:24Z) - Towards Scalable Multi-View Reconstruction of Geometry and Materials [27.660389147094715]
We propose a novel method for joint recovery of camera pose, object geometry and spatially-varying Bidirectional Reflectance Distribution Function (svBRDF) of 3D scenes.
The input are high-resolution RGBD images captured by a mobile, hand-held capture system with point lights for active illumination.
arXiv Detail & Related papers (2023-06-06T15:07:39Z) - DETR4D: Direct Multi-View 3D Object Detection with Sparse Attention [50.11672196146829]
3D object detection with surround-view images is an essential task for autonomous driving.
We propose DETR4D, a Transformer-based framework that explores sparse attention and direct feature query for 3D object detection in multi-view images.
arXiv Detail & Related papers (2022-12-15T14:18:47Z) - Leveraging Spatial and Photometric Context for Calibrated Non-Lambertian
Photometric Stereo [61.6260594326246]
We introduce an efficient fully-convolutional architecture that can leverage both spatial and photometric context simultaneously.
Using separable 4D convolutions and 2D heat-maps reduces the size and makes more efficient.
arXiv Detail & Related papers (2021-03-22T18:06:58Z) - Multi-view Depth Estimation using Epipolar Spatio-Temporal Networks [87.50632573601283]
We present a novel method for multi-view depth estimation from a single video.
Our method achieves temporally coherent depth estimation results by using a novel Epipolar Spatio-Temporal (EST) transformer.
To reduce the computational cost, inspired by recent Mixture-of-Experts models, we design a compact hybrid network.
arXiv Detail & Related papers (2020-11-26T04:04:21Z) - Deep 3D Capture: Geometry and Reflectance from Sparse Multi-View Images [59.906948203578544]
We introduce a novel learning-based method to reconstruct the high-quality geometry and complex, spatially-varying BRDF of an arbitrary object.
We first estimate per-view depth maps using a deep multi-view stereo network.
These depth maps are used to coarsely align the different views.
We propose a novel multi-view reflectance estimation network architecture.
arXiv Detail & Related papers (2020-03-27T21:28:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.