ConDor: Self-Supervised Canonicalization of 3D Pose for Partial Shapes
- URL: http://arxiv.org/abs/2201.07788v1
- Date: Wed, 19 Jan 2022 18:57:21 GMT
- Title: ConDor: Self-Supervised Canonicalization of 3D Pose for Partial Shapes
- Authors: Rahul Sajnani, Adrien Poulenard, Jivitesh Jain, Radhika Dua, Leonidas
J. Guibas, Srinath Sridhar
- Abstract summary: ConDor is a self-supervised method that learns to canonicalize the 3D orientation and position for full and partial 3D point clouds.
During inference, our method takes an unseen full or partial 3D point cloud at an arbitrary pose and outputs an equivariant canonical pose.
- Score: 55.689763519293464
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Progress in 3D object understanding has relied on manually canonicalized
shape datasets that contain instances with consistent position and orientation
(3D pose). This has made it hard to generalize these methods to in-the-wild
shapes, eg., from internet model collections or depth sensors. ConDor is a
self-supervised method that learns to Canonicalize the 3D orientation and
position for full and partial 3D point clouds. We build on top of Tensor Field
Networks (TFNs), a class of permutation- and rotation-equivariant, and
translation-invariant 3D networks. During inference, our method takes an unseen
full or partial 3D point cloud at an arbitrary pose and outputs an equivariant
canonical pose. During training, this network uses self-supervision losses to
learn the canonical pose from an un-canonicalized collection of full and
partial 3D point clouds. ConDor can also learn to consistently co-segment
object parts without any supervision. Extensive quantitative results on four
new metrics show that our approach outperforms existing methods while enabling
new applications such as operation on depth images and annotation transfer.
Related papers
- U3DS$^3$: Unsupervised 3D Semantic Scene Segmentation [19.706172244951116]
This paper presents U3DS$3$, as a step towards completely unsupervised point cloud segmentation for any holistic 3D scenes.
The initial step of our proposed approach involves generating superpoints based on the geometric characteristics of each scene.
We then undergo a learning process through a spatial clustering-based methodology, followed by iterative training using pseudo-labels generated in accordance with the cluster centroids.
arXiv Detail & Related papers (2023-11-10T12:05:35Z) - 3D Surface Reconstruction in the Wild by Deforming Shape Priors from
Synthetic Data [24.97027425606138]
Reconstructing the underlying 3D surface of an object from a single image is a challenging problem.
We present a new method for joint category-specific 3D reconstruction and object pose estimation from a single image.
Our approach achieves state-of-the-art reconstruction performance across several real-world datasets.
arXiv Detail & Related papers (2023-02-24T20:37:27Z) - End-to-End Learning of Multi-category 3D Pose and Shape Estimation [128.881857704338]
We propose an end-to-end method that simultaneously detects 2D keypoints from an image and lifts them to 3D.
The proposed method learns both 2D detection and 3D lifting only from 2D keypoints annotations.
In addition to being end-to-end in image to 3D learning, our method also handles objects from multiple categories using a single neural network.
arXiv Detail & Related papers (2021-12-19T17:10:40Z) - Learning Canonical 3D Object Representation for Fine-Grained Recognition [77.33501114409036]
We propose a novel framework for fine-grained object recognition that learns to recover object variation in 3D space from a single image.
We represent an object as a composition of 3D shape and its appearance, while eliminating the effect of camera viewpoint.
By incorporating 3D shape and appearance jointly in a deep representation, our method learns the discriminative representation of the object.
arXiv Detail & Related papers (2021-08-10T12:19:34Z) - ParaNet: Deep Regular Representation for 3D Point Clouds [62.81379889095186]
ParaNet is a novel end-to-end deep learning framework for representing 3D point clouds.
It converts an irregular 3D point cloud into a regular 2D color image, named point geometry image (PGI)
In contrast to conventional regular representation modalities based on multi-view projection and voxelization, the proposed representation is differentiable and reversible.
arXiv Detail & Related papers (2020-12-05T13:19:55Z) - Canonical 3D Deformer Maps: Unifying parametric and non-parametric
methods for dense weakly-supervised category reconstruction [79.98689027127855]
We propose a new representation of the 3D shape of common object categories that can be learned from a collection of 2D images of independent objects.
Our method builds in a novel way on concepts from parametric deformation models, non-parametric 3D reconstruction, and canonical embeddings.
It achieves state-of-the-art results in dense 3D reconstruction on public in-the-wild datasets of faces, cars, and birds.
arXiv Detail & Related papers (2020-08-28T15:44:05Z) - Weakly-supervised 3D Shape Completion in the Wild [91.04095516680438]
We address the problem of learning 3D complete shape from unaligned and real-world partial point clouds.
We propose a weakly-supervised method to estimate both 3D canonical shape and 6-DoF pose for alignment, given multiple partial observations.
Experiments on both synthetic and real data show that it is feasible and promising to learn 3D shape completion through large-scale data without shape and pose supervision.
arXiv Detail & Related papers (2020-08-20T17:53:42Z) - Learning Canonical Shape Space for Category-Level 6D Object Pose and
Size Estimation [21.7030393344051]
We learn canonical shape space (CASS), a unified representation for a large variety of instances of a certain object category.
We train a variational auto-encoder (VAE) for generating 3D point clouds in the canonical space from an RGBD image.
VAE is trained in a cross-category fashion, exploiting the publicly available large 3D shape repositories.
arXiv Detail & Related papers (2020-01-25T14:16:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.