Learning Implicit Probability Distribution Functions for Symmetric
Orientation Estimation from RGB Images Without Pose Labels
- URL: http://arxiv.org/abs/2211.11394v1
- Date: Mon, 21 Nov 2022 12:07:40 GMT
- Title: Learning Implicit Probability Distribution Functions for Symmetric
Orientation Estimation from RGB Images Without Pose Labels
- Authors: Arul Selvam Periyasamy, Luis Denninger, and Sven Behnke
- Abstract summary: We propose an automatic pose labeling scheme for RGB-D images.
We train an ImplicitPDF model to estimate the likelihood of an orientation hypothesis given an RGB image.
An efficient hierarchical sampling of the SO(3) manifold enables tractable generation of the complete set of symmetries.
- Score: 23.01797447932351
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Object pose estimation is a necessary prerequisite for autonomous robotic
manipulation, but the presence of symmetry increases the complexity of the pose
estimation task. Existing methods for object pose estimation output a single 6D
pose. Thus, they lack the ability to reason about symmetries. Lately, modeling
object orientation as a non-parametric probability distribution on the SO(3)
manifold by neural networks has shown impressive results. However, acquiring
large-scale datasets to train pose estimation models remains a bottleneck. To
address this limitation, we introduce an automatic pose labeling scheme. Given
RGB-D images without object pose annotations and 3D object models, we design a
two-stage pipeline consisting of point cloud registration and
render-and-compare validation to generate multiple symmetrical
pseudo-ground-truth pose labels for each image. Using the generated pose
labels, we train an ImplicitPDF model to estimate the likelihood of an
orientation hypothesis given an RGB image. An efficient hierarchical sampling
of the SO(3) manifold enables tractable generation of the complete set of
symmetries at multiple resolutions. During inference, the most likely
orientation of the target object is estimated using gradient ascent. We
evaluate the proposed automatic pose labeling scheme and the ImplicitPDF model
on a photorealistic dataset and the T-Less dataset, demonstrating the
advantages of the proposed method.
Related papers
- DVMNet: Computing Relative Pose for Unseen Objects Beyond Hypotheses [59.51874686414509]
Current approaches approximate the continuous pose representation with a large number of discrete pose hypotheses.
We present a Deep Voxel Matching Network (DVMNet) that eliminates the need for pose hypotheses and computes the relative object pose in a single pass.
Our method delivers more accurate relative pose estimates for novel objects at a lower computational cost compared to state-of-the-art methods.
arXiv Detail & Related papers (2024-03-20T15:41:32Z) - FoundPose: Unseen Object Pose Estimation with Foundation Features [11.32559845631345]
FoundPose is a model-based method for 6D pose estimation of unseen objects from a single RGB image.
The method can quickly onboard new objects using their 3D models without requiring any object- or task-specific training.
arXiv Detail & Related papers (2023-11-30T18:52:29Z) - Image to Sphere: Learning Equivariant Features for Efficient Pose
Prediction [3.823356975862006]
Methods that predict a single point estimate do not predict the pose of objects with symmetries well and cannot represent uncertainty.
We propose a novel mapping of features from the image domain to the 3D rotation manifold.
We demonstrate the effectiveness of our method at object orientation prediction, and achieve state-of-the-art performance on the popular PASCAL3D+ dataset.
arXiv Detail & Related papers (2023-02-27T16:23:19Z) - Semantic keypoint-based pose estimation from single RGB frames [64.80395521735463]
We present an approach to estimating the continuous 6-DoF pose of an object from a single RGB image.
The approach combines semantic keypoints predicted by a convolutional network (convnet) with a deformable shape model.
We show that our approach can accurately recover the 6-DoF object pose for both instance- and class-based scenarios.
arXiv Detail & Related papers (2022-04-12T15:03:51Z) - Uncertainty-Aware Adaptation for Self-Supervised 3D Human Pose
Estimation [70.32536356351706]
We introduce MRP-Net that constitutes a common deep network backbone with two output heads subscribing to two diverse configurations.
We derive suitable measures to quantify prediction uncertainty at both pose and joint level.
We present a comprehensive evaluation of the proposed approach and demonstrate state-of-the-art performance on benchmark datasets.
arXiv Detail & Related papers (2022-03-29T07:14:58Z) - Adversarial Parametric Pose Prior [106.12437086990853]
We learn a prior that restricts the SMPL parameters to values that produce realistic poses via adversarial training.
We show that our learned prior covers the diversity of the real-data distribution, facilitates optimization for 3D reconstruction from 2D keypoints, and yields better pose estimates when used for regression from images.
arXiv Detail & Related papers (2021-12-08T10:05:32Z) - Stochastic Modeling for Learnable Human Pose Triangulation [0.7646713951724009]
We propose a modeling framework for 3D human pose triangulation and evaluate its performance across different datasets and spatial camera arrangements.
The proposed pose triangulation model successfully generalizes to different camera arrangements and between two public datasets.
arXiv Detail & Related papers (2021-10-01T09:26:25Z) - Implicit-PDF: Non-Parametric Representation of Probability Distributions
on the Rotation Manifold [47.31074799708132]
We introduce a method to estimate arbitrary, non-parametric distributions on SO(3).
Our key idea is to represent the distributions implicitly, with a neural network that estimates the probability given the input image and a candidate pose.
We achieve state-of-the-art performance on Pascal3D+ and ModelNet10-SO(3) benchmarks.
arXiv Detail & Related papers (2021-06-10T17:57:23Z) - SymmetryNet: Learning to Predict Reflectional and Rotational Symmetries
of 3D Shapes from Single-View RGB-D Images [26.38270361331076]
We propose an end-to-end deep neural network which is able to predict both reflectional and rotational symmetries of 3D objects.
We also contribute a benchmark of 3D symmetry detection based on single-view RGB-D images.
arXiv Detail & Related papers (2020-08-02T14:10:09Z) - EPOS: Estimating 6D Pose of Objects with Symmetries [57.448933686429825]
We present a new method for estimating the 6D pose of rigid objects with available 3D models from a single RGB input.
An object is represented by compact surface fragments which allow symmetries in a systematic manner.
Correspondences between densely sampled pixels and the fragments are predicted using an encoder-decoder network.
arXiv Detail & Related papers (2020-04-01T17:41:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.