Physics to the Rescue: Deep Non-line-of-sight Reconstruction for
High-speed Imaging
- URL: http://arxiv.org/abs/2205.01679v1
- Date: Tue, 3 May 2022 02:47:02 GMT
- Title: Physics to the Rescue: Deep Non-line-of-sight Reconstruction for
High-speed Imaging
- Authors: Fangzhou Mu, Sicheng Mo, Jiayong Peng, Xiaochun Liu, Ji Hyun Nam,
Siddeshwar Raghavan, Andreas Velten and Yin Li
- Abstract summary: We present a novel deep model that incorporates the complementary physics priors of wave propagation and volume rendering into a neural network for high-quality and robust NLOS reconstruction.
Our method outperforms prior physics and learning based approaches on both synthetic and real measurements.
- Score: 13.271762773872476
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Computational approach to imaging around the corner, or non-line-of-sight
(NLOS) imaging, is becoming a reality thanks to major advances in imaging
hardware and reconstruction algorithms. A recent development towards practical
NLOS imaging, Nam et al. demonstrated a high-speed non-confocal imaging system
that operates at 5Hz, 100x faster than the prior art. This enormous gain in
acquisition rate, however, necessitates numerous approximations in light
transport, breaking many existing NLOS reconstruction methods that assume an
idealized image formation model. To bridge the gap, we present a novel deep
model that incorporates the complementary physics priors of wave propagation
and volume rendering into a neural network for high-quality and robust NLOS
reconstruction. This orchestrated design regularizes the solution space by
relaxing the image formation model, resulting in a deep model that generalizes
well on real captures despite being exclusively trained on synthetic data.
Further, we devise a unified learning framework that enables our model to be
flexibly trained using diverse supervision signals, including target intensity
images or even raw NLOS transient measurements. Once trained, our model renders
both intensity and depth images at inference time in a single forward pass,
capable of processing more than 5 captures per second on a high-end GPU.
Through extensive qualitative and quantitative experiments, we show that our
method outperforms prior physics and learning based approaches on both
synthetic and real measurements. We anticipate that our method along with the
fast capturing system will accelerate future development of NLOS imaging for
real world applications that require high-speed imaging.
Related papers
- Optical training of large-scale Transformers and deep neural networks with direct feedback alignment [48.90869997343841]
We experimentally implement a versatile and scalable training algorithm, called direct feedback alignment, on a hybrid electronic-photonic platform.
An optical processing unit performs large-scale random matrix multiplications, which is the central operation of this algorithm, at speeds up to 1500 TeraOps.
We study the compute scaling of our hybrid optical approach, and demonstrate a potential advantage for ultra-deep and wide neural networks.
arXiv Detail & Related papers (2024-09-01T12:48:47Z) - Efficient-3DiM: Learning a Generalizable Single-image Novel-view
Synthesizer in One Day [63.96075838322437]
We propose a framework to learn a single-image novel-view synthesizer.
Our framework is able to reduce the total training time from 10 days to less than 1 day.
arXiv Detail & Related papers (2023-10-04T17:57:07Z) - Physics-Driven Turbulence Image Restoration with Stochastic Refinement [80.79900297089176]
Image distortion by atmospheric turbulence is a critical problem in long-range optical imaging systems.
Fast and physics-grounded simulation tools have been introduced to help the deep-learning models adapt to real-world turbulence conditions.
This paper proposes the Physics-integrated Restoration Network (PiRN) to help the network to disentangle theity from the degradation and the underlying image.
arXiv Detail & Related papers (2023-07-20T05:49:21Z) - Physics Constrained Unsupervised Deep Learning for Rapid, High
Resolution Scanning Coherent Diffraction Reconstruction [0.4310181932415864]
Coherent diffractive imaging (CDI) and ptychography are making their way into scientific fields ranging from X-ray imaging to astronomy.
We propose an unsupervised physics-informed neural network reconstruction method, PtychoPINN.
In particular, PtychoPINN significantly advances generalizability, accuracy (with a typical 10 dB PSNR increase), and linear resolution (2- to 6-fold gain)
arXiv Detail & Related papers (2023-06-19T15:24:07Z) - Neural 3D Reconstruction in the Wild [86.6264706256377]
We introduce a new method that enables efficient and accurate surface reconstruction from Internet photo collections.
We present a new benchmark and protocol for evaluating reconstruction performance on such in-the-wild scenes.
arXiv Detail & Related papers (2022-05-25T17:59:53Z) - DH-GAN: A Physics-driven Untrained Generative Adversarial Network for 3D
Microscopic Imaging using Digital Holography [3.4635026053111484]
Digital holography is a 3D imaging technique by emitting a laser beam with a plane wavefront to an object and measuring the intensity of the diffracted waveform, called holograms.
Recently, deep learning (DL) methods have been used for more accurate holographic processing.
We propose a new DL architecture based on generative adversarial networks that uses a discriminative network for realizing a semantic measure for reconstruction quality.
arXiv Detail & Related papers (2022-05-25T17:13:45Z) - Adaptive Gradient Balancing for UndersampledMRI Reconstruction and
Image-to-Image Translation [60.663499381212425]
We enhance the image quality by using a Wasserstein Generative Adversarial Network combined with a novel Adaptive Gradient Balancing technique.
In MRI, our method minimizes artifacts, while maintaining a high-quality reconstruction that produces sharper images than other techniques.
arXiv Detail & Related papers (2021-04-05T13:05:22Z) - Back to Event Basics: Self-Supervised Learning of Image Reconstruction
for Event Cameras via Photometric Constancy [0.0]
Event cameras are novel vision sensors that sample, in an asynchronous fashion, brightness increments with low latency and high temporal resolution.
We propose a novel, lightweight neural network for optical flow estimation that achieves high speed inference with only a minor drop in performance.
Results across multiple datasets show that the performance of the proposed self-supervised approach is in line with the state-of-the-art.
arXiv Detail & Related papers (2020-09-17T13:30:05Z) - Two-shot Spatially-varying BRDF and Shape Estimation [89.29020624201708]
We propose a novel deep learning architecture with a stage-wise estimation of shape and SVBRDF.
We create a large-scale synthetic training dataset with domain-randomized geometry and realistic materials.
Experiments on both synthetic and real-world datasets show that our network trained on a synthetic dataset can generalize well to real-world images.
arXiv Detail & Related papers (2020-04-01T12:56:13Z) - Deep Non-Line-of-Sight Reconstruction [18.38481917675749]
In this paper, we employ convolutional feed-forward networks for solving the reconstruction problem efficiently.
We devise a tailored autoencoder architecture, trained end-to-end reconstruction maps transient images directly to a depth map representation.
We demonstrate that our feed-forward network, even though it is trained solely on synthetic data, generalizes to measured data from SPAD sensors and is able to obtain results that are competitive with model-based reconstruction methods.
arXiv Detail & Related papers (2020-01-24T16:05:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.