Reconstruct, Rasterize and Backprop: Dense shape and pose estimation
from a single image
- URL: http://arxiv.org/abs/2004.12232v1
- Date: Sat, 25 Apr 2020 20:53:43 GMT
- Title: Reconstruct, Rasterize and Backprop: Dense shape and pose estimation
from a single image
- Authors: Aniket Pokale, Aditya Aggarwal, K. Madhava Krishna
- Abstract summary: This paper presents a new system to obtain dense object reconstructions along with 6-DoF poses from a single image.
We leverage recent advances in differentiable rendering (in particular, robotics) to close the loop with 3D reconstruction in camera frame.
- Score: 14.9851111159799
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper presents a new system to obtain dense object reconstructions along
with 6-DoF poses from a single image. Geared towards high fidelity
reconstruction, several recent approaches leverage implicit surface
representations and deep neural networks to estimate a 3D mesh of an object,
given a single image. However, all such approaches recover only the shape of an
object; the reconstruction is often in a canonical frame, unsuitable for
downstream robotics tasks. To this end, we leverage recent advances in
differentiable rendering (in particular, rasterization) to close the loop with
3D reconstruction in camera frame. We demonstrate that our approach---dubbed
reconstruct, rasterize and backprop (RRB) achieves significantly lower pose
estimation errors compared to prior art, and is able to recover dense object
shapes and poses from imagery. We further extend our results to an (offline)
setup, where we demonstrate a dense monocular object-centric egomotion
estimation system.
Related papers
- Floating No More: Object-Ground Reconstruction from a Single Image [33.34421517827975]
We introduce ORG (Object Reconstruction with Ground), a novel task aimed at reconstructing 3D object geometry in conjunction with the ground surface.
Our method uses two compact pixel-level representations to depict the relationship between camera, object, and ground.
arXiv Detail & Related papers (2024-07-26T17:59:56Z) - 2L3: Lifting Imperfect Generated 2D Images into Accurate 3D [16.66666619143761]
Multi-view (MV) 3D reconstruction is a promising solution to fuse generated MV images into consistent 3D objects.
However, the generated images usually suffer from inconsistent lighting, misaligned geometry, and sparse views, leading to poor reconstruction quality.
We present a novel 3D reconstruction framework that leverages intrinsic decomposition guidance, transient-mono prior guidance, and view augmentation to cope with the three issues.
arXiv Detail & Related papers (2024-01-29T02:30:31Z) - LIST: Learning Implicitly from Spatial Transformers for Single-View 3D
Reconstruction [5.107705550575662]
List is a novel neural architecture that leverages local and global image features to reconstruct geometric and topological structure of a 3D object from a single image.
We show the superiority of our model in reconstructing 3D objects from both synthetic and real-world images against the state of the art.
arXiv Detail & Related papers (2023-07-23T01:01:27Z) - Reconstructing Hand-Held Objects from Monocular Video [95.06750686508315]
This paper presents an approach that reconstructs a hand-held object from a monocular video.
In contrast to many recent methods that directly predict object geometry by a trained network, the proposed approach does not require any learned prior to the object.
arXiv Detail & Related papers (2022-11-30T09:14:58Z) - Towards Non-Line-of-Sight Photography [48.491977359971855]
Non-line-of-sight (NLOS) imaging is based on capturing the multi-bounce indirect reflections from the hidden objects.
Active NLOS imaging systems rely on the capture of the time of flight of light through the scene.
We propose a new problem formulation, called NLOS photography, to specifically address this deficiency.
arXiv Detail & Related papers (2021-09-16T08:07:13Z) - Black-Box Test-Time Shape REFINEment for Single View 3D Reconstruction [57.805334118057665]
We propose REFINE, a postprocessing mesh refinement step that can be easily integrated into the pipeline of any black-box method in the literature.
At test time, REFINE optimize a network per mesh instance, to encourage consistency between the mesh and the given object view.
arXiv Detail & Related papers (2021-08-23T03:28:47Z) - Shape from Blur: Recovering Textured 3D Shape and Motion of Fast Moving
Objects [115.71874459429381]
We address the novel task of jointly reconstructing the 3D shape, texture, and motion of an object from a single motion-blurred image.
While previous approaches address the deblurring problem only in the 2D image domain, our proposed rigorous modeling of all object properties in the 3D domain enables the correct description of arbitrary object motion.
arXiv Detail & Related papers (2021-06-16T13:18:08Z) - From Points to Multi-Object 3D Reconstruction [71.17445805257196]
We propose a method to detect and reconstruct multiple 3D objects from a single RGB image.
A keypoint detector localizes objects as center points and directly predicts all object properties, including 9-DoF bounding boxes and 3D shapes.
The presented approach performs lightweight reconstruction in a single-stage, it is real-time capable, fully differentiable and end-to-end trainable.
arXiv Detail & Related papers (2020-12-21T18:52:21Z) - A Divide et Impera Approach for 3D Shape Reconstruction from Multiple
Views [49.03830902235915]
Estimating the 3D shape of an object from a single or multiple images has gained popularity thanks to the recent breakthroughs powered by deep learning.
This paper proposes to rely on viewpoint variant reconstructions by merging the visible information from the given views.
To validate the proposed method, we perform a comprehensive evaluation on the ShapeNet reference benchmark in terms of relative pose estimation and 3D shape reconstruction.
arXiv Detail & Related papers (2020-11-17T09:59:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.