NoPose-NeuS: Jointly Optimizing Camera Poses with Neural Implicit
Surfaces for Multi-view Reconstruction
- URL: http://arxiv.org/abs/2312.15238v1
- Date: Sat, 23 Dec 2023 12:18:22 GMT
- Title: NoPose-NeuS: Jointly Optimizing Camera Poses with Neural Implicit
Surfaces for Multi-view Reconstruction
- Authors: Mohamed Shawky Sabae, Hoda Anis Baraka, Mayada Mansour Hadhoud
- Abstract summary: NoPose-NeuS is a neural implicit surface reconstruction method that extends NeuS to jointly optimize camera poses with the geometry and color networks.
We show that the proposed method can estimate relatively accurate camera poses, while maintaining a high surface reconstruction quality with 0.89 mean Chamfer distance.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Learning neural implicit surfaces from volume rendering has become popular
for multi-view reconstruction. Neural surface reconstruction approaches can
recover complex 3D geometry that are difficult for classical Multi-view Stereo
(MVS) approaches, such as non-Lambertian surfaces and thin structures. However,
one key assumption for these methods is knowing accurate camera parameters for
the input multi-view images, which are not always available. In this paper, we
present NoPose-NeuS, a neural implicit surface reconstruction method that
extends NeuS to jointly optimize camera poses with the geometry and color
networks. We encode the camera poses as a multi-layer perceptron (MLP) and
introduce two additional losses, which are multi-view feature consistency and
rendered depth losses, to constrain the learned geometry for better estimated
camera poses and scene surfaces. Extensive experiments on the DTU dataset show
that the proposed method can estimate relatively accurate camera poses, while
maintaining a high surface reconstruction quality with 0.89 mean Chamfer
distance.
Related papers
- NeRSP: Neural 3D Reconstruction for Reflective Objects with Sparse Polarized Images [62.752710734332894]
NeRSP is a Neural 3D reconstruction technique for Reflective surfaces with Sparse Polarized images.
We derive photometric and geometric cues from the polarimetric image formation model and multiview azimuth consistency.
We achieve the state-of-the-art surface reconstruction results with only 6 views as input.
arXiv Detail & Related papers (2024-06-11T09:53:18Z) - FrozenRecon: Pose-free 3D Scene Reconstruction with Frozen Depth Models [67.96827539201071]
We propose a novel test-time optimization approach for 3D scene reconstruction.
Our method achieves state-of-the-art cross-dataset reconstruction on five zero-shot testing datasets.
arXiv Detail & Related papers (2023-08-10T17:55:02Z) - SC-NeuS: Consistent Neural Surface Reconstruction from Sparse and Noisy
Views [20.840876921128956]
This paper pays special attention on the consistent surface reconstruction from sparse views with noisy camera poses.
Unlike previous approaches, the key difference of this paper is to exploit the multi-view constraints directly from the explicit geometry of the neural surface.
We propose a jointly learning strategy for neural surface and camera poses, named SC-NeuS, to perform geometry-consistent surface reconstruction in an end-to-end manner.
arXiv Detail & Related papers (2023-07-12T03:45:45Z) - Level-S$^2$fM: Structure from Motion on Neural Level Set of Implicit
Surfaces [36.06713735409501]
This paper presents a neural incremental Structure-from-Motion (SfM) approach, Level-S$2$fM, which estimates the camera poses and scene geometry from a set of uncalibrated images.
Our novel formulation poses some new challenges due to inevitable two-view and few-view configurations in the incremental SfM pipeline.
Not only does our Level-S$2$fM lead to promising results on camera pose estimation and scene geometry reconstruction, but it also shows a promising way for neural implicit rendering without knowing camera.
arXiv Detail & Related papers (2022-11-22T05:21:21Z) - Multi-View Neural Surface Reconstruction with Structured Light [7.709526244898887]
Three-dimensional (3D) object reconstruction based on differentiable rendering (DR) is an active research topic in computer vision.
We introduce active sensing with structured light (SL) into multi-view 3D object reconstruction based on DR to learn the unknown geometry and appearance of arbitrary scenes and camera poses.
Our method realizes high reconstruction accuracy in the textureless region and reduces efforts for camera pose calibration.
arXiv Detail & Related papers (2022-11-22T03:10:46Z) - MonoSDF: Exploring Monocular Geometric Cues for Neural Implicit Surface
Reconstruction [72.05649682685197]
State-of-the-art neural implicit methods allow for high-quality reconstructions of simple scenes from many input views.
This is caused primarily by the inherent ambiguity in the RGB reconstruction loss that does not provide enough constraints.
Motivated by recent advances in the area of monocular geometry prediction, we explore the utility these cues provide for improving neural implicit surface reconstruction.
arXiv Detail & Related papers (2022-06-01T17:58:15Z) - Towards Non-Line-of-Sight Photography [48.491977359971855]
Non-line-of-sight (NLOS) imaging is based on capturing the multi-bounce indirect reflections from the hidden objects.
Active NLOS imaging systems rely on the capture of the time of flight of light through the scene.
We propose a new problem formulation, called NLOS photography, to specifically address this deficiency.
arXiv Detail & Related papers (2021-09-16T08:07:13Z) - Vid2Curve: Simultaneous Camera Motion Estimation and Thin Structure
Reconstruction from an RGB Video [90.93141123721713]
Thin structures, such as wire-frame sculptures, fences, cables, power lines, and tree branches, are common in the real world.
It is extremely challenging to acquire their 3D digital models using traditional image-based or depth-based reconstruction methods because thin structures often lack distinct point features and have severe self-occlusion.
We propose the first approach that simultaneously estimates camera motion and reconstructs the geometry of complex 3D thin structures in high quality from a color video captured by a handheld camera.
arXiv Detail & Related papers (2020-05-07T10:39:20Z) - Deep 3D Capture: Geometry and Reflectance from Sparse Multi-View Images [59.906948203578544]
We introduce a novel learning-based method to reconstruct the high-quality geometry and complex, spatially-varying BRDF of an arbitrary object.
We first estimate per-view depth maps using a deep multi-view stereo network.
These depth maps are used to coarsely align the different views.
We propose a novel multi-view reflectance estimation network architecture.
arXiv Detail & Related papers (2020-03-27T21:28:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.