Reconstructing Small 3D Objects in front of a Textured Background
- URL: http://arxiv.org/abs/2105.11352v1
- Date: Mon, 24 May 2021 15:36:33 GMT
- Title: Reconstructing Small 3D Objects in front of a Textured Background
- Authors: Petr Hruby and Tomas Pajdla
- Abstract summary: We present a technique for a complete 3D reconstruction of small objects moving in front of a textured background.
It is a particular variation of multibody structure from motion, which specializes to two objects only.
In experiments with real artifacts, we show that our approach has practical advantages when reconstructing 3D objects from all sides.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We present a technique for a complete 3D reconstruction of small objects
moving in front of a textured background. It is a particular variation of
multibody structure from motion, which specializes to two objects only. The
scene is captured in several static configurations between which the relative
pose of the two objects may change. We reconstruct every static configuration
individually and segment the points locally by finding multiple poses of
cameras that capture the scene's other configurations. Then, the local
segmentation results are combined, and the reconstructions are merged into the
resulting model of the scene. In experiments with real artifacts, we show that
our approach has practical advantages when reconstructing 3D objects from all
sides. In this setting, our method outperforms the state-of-the-art. We
integrate our method into the state of the art 3D reconstruction pipeline
COLMAP.
Related papers
- Floating No More: Object-Ground Reconstruction from a Single Image [33.34421517827975]
We introduce ORG (Object Reconstruction with Ground), a novel task aimed at reconstructing 3D object geometry in conjunction with the ground surface.
Our method uses two compact pixel-level representations to depict the relationship between camera, object, and ground.
arXiv Detail & Related papers (2024-07-26T17:59:56Z) - Part123: Part-aware 3D Reconstruction from a Single-view Image [54.589723979757515]
Part123 is a novel framework for part-aware 3D reconstruction from a single-view image.
We introduce contrastive learning into a neural rendering framework to learn a part-aware feature space.
A clustering-based algorithm is also developed to automatically derive 3D part segmentation results from the reconstructed models.
arXiv Detail & Related papers (2024-05-27T07:10:21Z) - Disentangled 3D Scene Generation with Layout Learning [109.03233745767062]
We introduce a method to generate 3D scenes that are disentangled into their component objects.
Our key insight is that objects can be discovered by finding parts of a 3D scene that, when rearranged spatially, still produce valid configurations of the same scene.
We show that despite its simplicity, our approach successfully generates 3D scenes into individual objects.
arXiv Detail & Related papers (2024-02-26T18:54:15Z) - Towards High-Fidelity Single-view Holistic Reconstruction of Indoor
Scenes [50.317223783035075]
We present a new framework to reconstruct holistic 3D indoor scenes from single-view images.
We propose an instance-aligned implicit function (InstPIFu) for detailed object reconstruction.
Our code and model will be made publicly available.
arXiv Detail & Related papers (2022-07-18T14:54:57Z) - Deep Permutation Equivariant Structure from Motion [38.68492294795315]
Existing deep methods produce highly accurate 3D reconstructions in stereo and multiview stereo settings.
We propose a neural network architecture that recovers both the camera parameters and a scene structure by minimizing an unsupervised reprojection loss.
Our experiments, conducted on a variety of datasets in both internally calibrated and uncalibrated settings, indicate that our method accurately recovers pose and structure, on par with classical state of the art methods.
arXiv Detail & Related papers (2021-04-14T08:50:06Z) - A Divide et Impera Approach for 3D Shape Reconstruction from Multiple
Views [49.03830902235915]
Estimating the 3D shape of an object from a single or multiple images has gained popularity thanks to the recent breakthroughs powered by deep learning.
This paper proposes to rely on viewpoint variant reconstructions by merging the visible information from the given views.
To validate the proposed method, we perform a comprehensive evaluation on the ShapeNet reference benchmark in terms of relative pose estimation and 3D shape reconstruction.
arXiv Detail & Related papers (2020-11-17T09:59:32Z) - AutoSweep: Recovering 3D Editable Objectsfrom a Single Photograph [54.701098964773756]
We aim to recover 3D objects with semantic parts and can be directly edited.
Our work makes an attempt towards recovering two types of primitive-shaped objects, namely, generalized cuboids and generalized cylinders.
Our algorithm can recover high quality 3D models and outperforms existing methods in both instance segmentation and 3D reconstruction.
arXiv Detail & Related papers (2020-05-27T12:16:24Z) - CoReNet: Coherent 3D scene reconstruction from a single RGB image [43.74240268086773]
We build on advances in deep learning to reconstruct the shape of a single object given only one RBG image as input.
We propose three extensions: (1) ray-traced skip connections that propagate local 2D information to the output 3D volume in a physically correct manner; (2) a hybrid 3D volume representation that enables building translation equivariant models; and (3) a reconstruction loss tailored to capture overall object geometry.
We reconstruct all objects jointly in one pass, producing a coherent reconstruction, where all objects live in a single consistent 3D coordinate frame relative to the camera and they do not intersect in 3D space.
arXiv Detail & Related papers (2020-04-27T17:53:07Z) - Self-supervised Single-view 3D Reconstruction via Semantic Consistency [142.71430568330172]
We learn a self-supervised, single-view 3D reconstruction model that predicts the shape, texture and camera pose of a target object.
The proposed method does not necessitate 3D supervision, manually annotated keypoints, multi-view images of an object or a prior 3D template.
arXiv Detail & Related papers (2020-03-13T20:29:01Z) - Total3DUnderstanding: Joint Layout, Object Pose and Mesh Reconstruction
for Indoor Scenes from a Single Image [24.99186733297264]
We propose an end-to-end solution to jointly reconstruct room layout, object bounding boxes and meshes from a single image.
Our method builds upon a holistic scene context and proposes a coarse-to-fine hierarchy with three components.
Experiments on the SUN RGB-D and Pix3D datasets demonstrate that our method consistently outperforms existing methods.
arXiv Detail & Related papers (2020-02-27T16:00:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.