NormNet: Scale Normalization for 6D Pose Estimation in Stacked Scenarios
- URL: http://arxiv.org/abs/2311.09269v1
- Date: Wed, 15 Nov 2023 12:02:57 GMT
- Title: NormNet: Scale Normalization for 6D Pose Estimation in Stacked Scenarios
- Authors: En-Te Lin, Wei-Jie Lv, Ding-Tao Huang and Long Zeng
- Abstract summary: This paper proposes a new 6DoF OPE network (NormNet) for different scale objects in stacked scenarios.
All objects in the stacked scenario are normalized into the same scale through semantic segmentation and affine transformation.
Finally, they are fed into a shared pose estimator to recover their 6D poses.
- Score: 4.515332570030772
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Existing Object Pose Estimation (OPE) methods for stacked scenarios are not
robust to changes in object scale. This paper proposes a new 6DoF OPE network
(NormNet) for different scale objects in stacked scenarios. Specifically, each
object's scale is first learned with point-wise regression. Then, all objects
in the stacked scenario are normalized into the same scale through semantic
segmentation and affine transformation. Finally, they are fed into a shared
pose estimator to recover their 6D poses. In addition, we introduce a new
Sim-to-Real transfer pipeline, combining style transfer and domain
randomization. This improves the NormNet's performance on real data even if we
only train it on synthetic data. Extensive experiments demonstrate that the
proposed method achieves state-of-the-art performance on public benchmarks and
the MultiScale dataset we constructed. The real-world experiments show that our
method can robustly estimate the 6D pose of objects at different scales.
Related papers
- FoundationPose: Unified 6D Pose Estimation and Tracking of Novel Objects [55.77542145604758]
FoundationPose is a unified foundation model for 6D object pose estimation and tracking.
Our approach can be instantly applied at test-time to a novel object without fine-tuning.
arXiv Detail & Related papers (2023-12-13T18:28:09Z) - Learning to Estimate 6DoF Pose from Limited Data: A Few-Shot,
Generalizable Approach using RGB Images [60.0898989456276]
We present a new framework named Cas6D for few-shot 6DoF pose estimation that is generalizable and uses only RGB images.
To address the false positives of target object detection in the extreme few-shot setting, our framework utilizes a self-supervised pre-trained ViT to learn robust feature representations.
Experimental results on the LINEMOD and GenMOP datasets demonstrate that Cas6D outperforms state-of-the-art methods by 9.2% and 3.8% accuracy (Proj-5) under the 32-shot setting.
arXiv Detail & Related papers (2023-06-13T07:45:42Z) - LocPoseNet: Robust Location Prior for Unseen Object Pose Estimation [69.70498875887611]
LocPoseNet is able to robustly learn location prior for unseen objects.
Our method outperforms existing works by a large margin on LINEMOD and GenMOP.
arXiv Detail & Related papers (2022-11-29T15:21:34Z) - Category-Level 6D Object Pose Estimation in the Wild: A Semi-Supervised
Learning Approach and A New Dataset [12.40437099476348]
We generalize category-level 6D object pose estimation in the wild with semi-supervised learning.
We propose a new model, called Rendering for Pose estimation network RePoNet, that is jointly trained using the free ground-truths with the synthetic data.
Our method outperforms state-of-the-art methods on the previous dataset and our Wild6D test set (with manual annotations for evaluation) by a large margin.
arXiv Detail & Related papers (2022-06-30T17:30:57Z) - Unseen Object 6D Pose Estimation: A Benchmark and Baselines [62.8809734237213]
We propose a new task that enables and facilitates algorithms to estimate the 6D pose estimation of novel objects during testing.
We collect a dataset with both real and synthetic images and up to 48 unseen objects in the test set.
By training an end-to-end 3D correspondences network, our method finds corresponding points between an unseen object and a partial view RGBD image accurately and efficiently.
arXiv Detail & Related papers (2022-06-23T16:29:53Z) - FS6D: Few-Shot 6D Pose Estimation of Novel Objects [116.34922994123973]
6D object pose estimation networks are limited in their capability to scale to large numbers of object instances.
In this work, we study a new open set problem; the few-shot 6D object poses estimation: estimating the 6D pose of an unknown object by a few support views without extra training.
arXiv Detail & Related papers (2022-03-28T10:31:29Z) - OVE6D: Object Viewpoint Encoding for Depth-based 6D Object Pose
Estimation [12.773040823634908]
We propose a universal framework, called OVE6D, for model-based 6D object pose estimation from a single depth image and a target object mask.
Our model is trained using purely synthetic data rendered from ShapeNet, and, unlike most of the existing methods, it generalizes well on new real-world objects without any fine-tuning.
We show that OVE6D outperforms some contemporary deep learning-based pose estimation methods specifically trained for individual objects or datasets with real-world training data.
arXiv Detail & Related papers (2022-03-02T12:51:33Z) - Shape Prior Deformation for Categorical 6D Object Pose and Size
Estimation [62.618227434286]
We present a novel learning approach to recover the 6D poses and sizes of unseen object instances from an RGB-D image.
We propose a deep network to reconstruct the 3D object model by explicitly modeling the deformation from a pre-learned categorical shape prior.
arXiv Detail & Related papers (2020-07-16T16:45:05Z) - Single Shot 6D Object Pose Estimation [11.37625512264302]
We introduce a novel single shot approach for 6D object pose estimation of rigid objects based on depth images.
A fully convolutional neural network is employed, where the 3D input data is spatially discretized and pose estimation is considered as a regression task.
With 65 fps on a GPU, our Object Pose Network (OP-Net) is extremely fast, is optimized end-to-end, and estimates the 6D pose of multiple objects in the image simultaneously.
arXiv Detail & Related papers (2020-04-27T11:59:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.