LEAP: Liberate Sparse-view 3D Modeling from Camera Poses
- URL: http://arxiv.org/abs/2310.01410v1
- Date: Mon, 2 Oct 2023 17:59:37 GMT
- Title: LEAP: Liberate Sparse-view 3D Modeling from Camera Poses
- Authors: Hanwen Jiang, Zhenyu Jiang, Yue Zhao, Qixing Huang
- Abstract summary: We present LEAP, a pose-free approach for sparse-view 3D modeling.
LEAP discards pose-based operations and learns geometric knowledge from data.
We show LEAP significantly outperforms prior methods when they employ predicted poses from state-of-the-art pose estimators.
- Score: 28.571234973474077
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Are camera poses necessary for multi-view 3D modeling? Existing approaches
predominantly assume access to accurate camera poses. While this assumption
might hold for dense views, accurately estimating camera poses for sparse views
is often elusive. Our analysis reveals that noisy estimated poses lead to
degraded performance for existing sparse-view 3D modeling methods. To address
this issue, we present LEAP, a novel pose-free approach, therefore challenging
the prevailing notion that camera poses are indispensable. LEAP discards
pose-based operations and learns geometric knowledge from data. LEAP is
equipped with a neural volume, which is shared across scenes and is
parameterized to encode geometry and texture priors. For each incoming scene,
we update the neural volume by aggregating 2D image features in a
feature-similarity-driven manner. The updated neural volume is decoded into the
radiance field, enabling novel view synthesis from any viewpoint. On both
object-centric and scene-level datasets, we show that LEAP significantly
outperforms prior methods when they employ predicted poses from
state-of-the-art pose estimators. Notably, LEAP performs on par with prior
approaches that use ground-truth poses while running $400\times$ faster than
PixelNeRF. We show LEAP generalizes to novel object categories and scenes, and
learns knowledge closely resembles epipolar geometry. Project page:
https://hwjiang1510.github.io/LEAP/
Related papers
- SpaRP: Fast 3D Object Reconstruction and Pose Estimation from Sparse Views [36.02533658048349]
We propose a novel method, SpaRP, to reconstruct a 3D textured mesh and estimate the relative camera poses for sparse-view images.
SpaRP distills knowledge from 2D diffusion models and finetunes them to implicitly deduce the 3D spatial relationships between the sparse views.
It requires only about 20 seconds to produce a textured mesh and camera poses for the input views.
arXiv Detail & Related papers (2024-08-19T17:53:10Z) - COLMAP-Free 3D Gaussian Splatting [88.420322646756]
We propose a novel method to perform novel view synthesis without any SfM preprocessing.
We process the input frames in a sequential manner and progressively grow the 3D Gaussians set by taking one input frame at a time.
Our method significantly improves over previous approaches in view synthesis and camera pose estimation under large motion changes.
arXiv Detail & Related papers (2023-12-12T18:39:52Z) - PF-LRM: Pose-Free Large Reconstruction Model for Joint Pose and Shape
Prediction [77.89935657608926]
We propose a Pose-Free Large Reconstruction Model (PF-LRM) for reconstructing a 3D object from a few unposed images.
PF-LRM simultaneously estimates the relative camera poses in 1.3 seconds on a single A100 GPU.
arXiv Detail & Related papers (2023-11-20T18:57:55Z) - VirtualPose: Learning Generalizable 3D Human Pose Models from Virtual
Data [69.64723752430244]
We introduce VirtualPose, a two-stage learning framework to exploit the hidden "free lunch" specific to this task.
The first stage transforms images to abstract geometry representations (AGR), and then the second maps them to 3D poses.
It addresses the generalization issue from two aspects: (1) the first stage can be trained on diverse 2D datasets to reduce the risk of over-fitting to limited appearance; (2) the second stage can be trained on diverse AGR synthesized from a large number of virtual cameras and poses.
arXiv Detail & Related papers (2022-07-20T14:47:28Z) - Object-Based Visual Camera Pose Estimation From Ellipsoidal Model and
3D-Aware Ellipse Prediction [2.016317500787292]
We propose a method for initial camera pose estimation from just a single image.
It exploits the ability of deep learning techniques to reliably detect objects regardless of viewing conditions.
Experiments prove that the accuracy of the computed pose significantly increases thanks to our method.
arXiv Detail & Related papers (2022-03-09T10:00:52Z) - MetaPose: Fast 3D Pose from Multiple Views without 3D Supervision [72.5863451123577]
We show how to train a neural model that can perform accurate 3D pose and camera estimation.
Our method outperforms both classical bundle adjustment and weakly-supervised monocular 3D baselines.
arXiv Detail & Related papers (2021-08-10T18:39:56Z) - CAMPARI: Camera-Aware Decomposed Generative Neural Radiance Fields [67.76151996543588]
We learn a 3D- and camera-aware generative model which faithfully recovers not only the image but also the camera data distribution.
At test time, our model generates images with explicit control over the camera as well as the shape and appearance of the scene.
arXiv Detail & Related papers (2021-03-31T17:59:24Z) - Back to the Feature: Learning Robust Camera Localization from Pixels to
Pose [114.89389528198738]
We introduce PixLoc, a scene-agnostic neural network that estimates an accurate 6-DoF pose from an image and a 3D model.
The system can localize in large environments given coarse pose priors but also improve the accuracy of sparse feature matching.
arXiv Detail & Related papers (2021-03-16T17:40:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.