CPPF: Towards Robust Category-Level 9D Pose Estimation in the Wild
- URL: http://arxiv.org/abs/2203.03089v1
- Date: Mon, 7 Mar 2022 01:36:22 GMT
- Title: CPPF: Towards Robust Category-Level 9D Pose Estimation in the Wild
- Authors: Yang You, Ruoxi Shi, Weiming Wang, Cewu Lu
- Abstract summary: Category-level PPF voting method to achieve accurate, robust and generalizable 9D pose estimation in the wild.
A novel coarse-to-fine voting algorithm is proposed to eliminate noisy point pair samples and generate final predictions from the population.
Our method is on par with current state of the arts with real-world training data.
- Score: 45.93626858034774
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we tackle the problem of category-level 9D pose estimation in
the wild, given a single RGB-D frame. Using supervised data of real-world 9D
poses is tedious and erroneous, and also fails to generalize to unseen
scenarios. Besides, category-level pose estimation requires a method to be able
to generalize to unseen objects at test time, which is also challenging.
Drawing inspirations from traditional point pair features (PPFs), in this
paper, we design a novel Category-level PPF (CPPF) voting method to achieve
accurate, robust and generalizable 9D pose estimation in the wild. To obtain
robust pose estimation, we sample numerous point pairs on an object, and for
each pair our model predicts necessary SE(3)-invariant voting statistics on
object centers, orientations and scales. A novel coarse-to-fine voting
algorithm is proposed to eliminate noisy point pair samples and generate final
predictions from the population. To get rid of false positives in the
orientation voting process, an auxiliary binary disambiguating classification
task is introduced for each sampled point pair. In order to detect objects in
the wild, we carefully design our sim-to-real pipeline by training on synthetic
point clouds only, unless objects have ambiguous poses in geometry. Under this
circumstance, color information is leveraged to disambiguate these poses.
Results on standard benchmarks show that our method is on par with current
state of the arts with real-world training data. Extensive experiments further
show that our method is robust to noise and gives promising results under
extremely challenging scenarios. Our code is available on
https://github.com/qq456cvb/CPPF.
Related papers
- LaPose: Laplacian Mixture Shape Modeling for RGB-Based Category-Level Object Pose Estimation [43.549593231397644]
LaPose is a novel framework that models the object shape as the Laplacian mixture model for Pose estimation.
By representing each point as a probabilistic distribution, we explicitly quantify the shape uncertainty.
LaPose yields state-of-the-art performance in category-level object pose estimation.
arXiv Detail & Related papers (2024-09-24T04:20:18Z) - DVMNet: Computing Relative Pose for Unseen Objects Beyond Hypotheses [59.51874686414509]
Current approaches approximate the continuous pose representation with a large number of discrete pose hypotheses.
We present a Deep Voxel Matching Network (DVMNet) that eliminates the need for pose hypotheses and computes the relative object pose in a single pass.
Our method delivers more accurate relative pose estimates for novel objects at a lower computational cost compared to state-of-the-art methods.
arXiv Detail & Related papers (2024-03-20T15:41:32Z) - PoseMatcher: One-shot 6D Object Pose Estimation by Deep Feature Matching [51.142988196855484]
We propose PoseMatcher, an accurate model free one-shot object pose estimator.
We create a new training pipeline for object to image matching based on a three-view system.
To enable PoseMatcher to attend to distinct input modalities, an image and a pointcloud, we introduce IO-Layer.
arXiv Detail & Related papers (2023-04-03T21:14:59Z) - CPPF++: Uncertainty-Aware Sim2Real Object Pose Estimation by Vote Aggregation [67.12857074801731]
We introduce a novel method, CPPF++, designed for sim-to-real pose estimation.
To address the challenge posed by vote collision, we propose a novel approach that involves modeling the voting uncertainty.
We incorporate several innovative modules, including noisy pair filtering, online alignment optimization, and a feature ensemble.
arXiv Detail & Related papers (2022-11-24T03:27:00Z) - Unseen Object 6D Pose Estimation: A Benchmark and Baselines [62.8809734237213]
We propose a new task that enables and facilitates algorithms to estimate the 6D pose estimation of novel objects during testing.
We collect a dataset with both real and synthetic images and up to 48 unseen objects in the test set.
By training an end-to-end 3D correspondences network, our method finds corresponding points between an unseen object and a partial view RGBD image accurately and efficiently.
arXiv Detail & Related papers (2022-06-23T16:29:53Z) - Zero-Shot Category-Level Object Pose Estimation [24.822189326540105]
We tackle the problem of estimating the pose of novel object categories in a zero-shot manner.
This extends much of the existing literature by removing the need for pose-labelled datasets or category-specific CAD models.
Our method provides a six-fold improvement in average rotation accuracy at 30 degrees.
arXiv Detail & Related papers (2022-04-07T17:58:39Z) - ZebraPose: Coarse to Fine Surface Encoding for 6DoF Object Pose
Estimation [76.31125154523056]
We present a discrete descriptor, which can represent the object surface densely.
We also propose a coarse to fine training strategy, which enables fine-grained correspondence prediction.
arXiv Detail & Related papers (2022-03-17T16:16:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.