StrobeNet: Category-Level Multiview Reconstruction of Articulated
Objects
- URL: http://arxiv.org/abs/2105.08016v1
- Date: Mon, 17 May 2021 17:05:42 GMT
- Title: StrobeNet: Category-Level Multiview Reconstruction of Articulated
Objects
- Authors: Ge Zhang, Or Litany, Srinath Sridhar, Leonidas Guibas
- Abstract summary: StrobeNet is a method for category-level 3D reconstruction of articulating objects from unposed RGB images.
Our approach reconstructs objects even when they are observed in different articulations in images with large baselines.
- Score: 17.698319441265223
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present StrobeNet, a method for category-level 3D reconstruction of
articulating objects from one or more unposed RGB images. Reconstructing
general articulating object categories % has important applications, but is
challenging since objects can have wide variation in shape, articulation,
appearance and topology. We address this by building on the idea of
category-level articulation canonicalization -- mapping observations to a
canonical articulation which enables correspondence-free multiview aggregation.
Our end-to-end trainable neural network estimates feature-enriched canonical 3D
point clouds, articulation joints, and part segmentation from one or more
unposed images of an object. These intermediate estimates are used to generate
a final implicit 3D reconstruction.Our approach reconstructs objects even when
they are observed in different articulations in images with large baselines,
and animation of reconstructed shapes. Quantitative and qualitative evaluations
on different object categories show that our method is able to achieve high
reconstruction accuracy, especially as more views are added.
Related papers
- Category-level Neural Field for Reconstruction of Partially Observed Objects in Indoor Environment [24.880495520422006]
We introduce category-level neural fields that learn meaningful common 3D information among objects belonging to the same category present in the scene.
Our key idea is to subcategorize objects based on their observed shape for better training of the category-level model.
Experiments on both simulation and real-world datasets demonstrate that our method improves the reconstruction of unobserved parts for several categories.
arXiv Detail & Related papers (2024-06-12T13:09:59Z) - Part123: Part-aware 3D Reconstruction from a Single-view Image [54.589723979757515]
Part123 is a novel framework for part-aware 3D reconstruction from a single-view image.
We introduce contrastive learning into a neural rendering framework to learn a part-aware feature space.
A clustering-based algorithm is also developed to automatically derive 3D part segmentation results from the reconstructed models.
arXiv Detail & Related papers (2024-05-27T07:10:21Z) - Single-view 3D Mesh Reconstruction for Seen and Unseen Categories [69.29406107513621]
Single-view 3D Mesh Reconstruction is a fundamental computer vision task that aims at recovering 3D shapes from single-view RGB images.
This paper tackles Single-view 3D Mesh Reconstruction, to study the model generalization on unseen categories.
We propose an end-to-end two-stage network, GenMesh, to break the category boundaries in reconstruction.
arXiv Detail & Related papers (2022-08-04T14:13:35Z) - Pre-train, Self-train, Distill: A simple recipe for Supersizing 3D
Reconstruction [47.38670633513938]
We learn a unified model for single-view 3D reconstruction of objects from hundreds of semantic categories.
Our work relies on segmented image collections for learning 3D of generic categories.
arXiv Detail & Related papers (2022-04-07T17:59:25Z) - Unsupervised Learning of 3D Object Categories from Videos in the Wild [75.09720013151247]
We focus on learning a model from multiple views of a large collection of object instances.
We propose a new neural network design, called warp-conditioned ray embedding (WCR), which significantly improves reconstruction.
Our evaluation demonstrates performance improvements over several deep monocular reconstruction baselines on existing benchmarks.
arXiv Detail & Related papers (2021-03-30T17:57:01Z) - Canonical 3D Deformer Maps: Unifying parametric and non-parametric
methods for dense weakly-supervised category reconstruction [79.98689027127855]
We propose a new representation of the 3D shape of common object categories that can be learned from a collection of 2D images of independent objects.
Our method builds in a novel way on concepts from parametric deformation models, non-parametric 3D reconstruction, and canonical embeddings.
It achieves state-of-the-art results in dense 3D reconstruction on public in-the-wild datasets of faces, cars, and birds.
arXiv Detail & Related papers (2020-08-28T15:44:05Z) - Learning Unsupervised Hierarchical Part Decomposition of 3D Objects from
a Single RGB Image [102.44347847154867]
We propose a novel formulation that allows to jointly recover the geometry of a 3D object as a set of primitives.
Our model recovers the higher level structural decomposition of various objects in the form of a binary tree of primitives.
Our experiments on the ShapeNet and D-FAUST datasets demonstrate that considering the organization of parts indeed facilitates reasoning about 3D geometry.
arXiv Detail & Related papers (2020-04-02T17:58:05Z) - Self-supervised Single-view 3D Reconstruction via Semantic Consistency [142.71430568330172]
We learn a self-supervised, single-view 3D reconstruction model that predicts the shape, texture and camera pose of a target object.
The proposed method does not necessitate 3D supervision, manually annotated keypoints, multi-view images of an object or a prior 3D template.
arXiv Detail & Related papers (2020-03-13T20:29:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.