Instance-specific 6-DoF Object Pose Estimation from Minimal Annotations
- URL: http://arxiv.org/abs/2207.13264v1
- Date: Wed, 27 Jul 2022 03:00:28 GMT
- Title: Instance-specific 6-DoF Object Pose Estimation from Minimal Annotations
- Authors: Rohan Pratap Singh, Iori Kumagai, Antonio Gabas, Mehdi Benallegue,
Yusuke Yoshiyasu, Fumio Kanehiro
- Abstract summary: We present a method to rapidly train and deploy a pipeline for estimating the continuous 6-DoF pose of an object from a single RGB image.
The key idea is to leverage known camera poses and rigid body geometry to partially automate the generation of a large labeled dataset.
The dataset, along with sufficient domain randomization, is then used to supervise the training of deep neural networks for predicting semantic keypoints.
- Score: 6.24717069374781
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In many robotic applications, the environment setting in which the 6-DoF pose
estimation of a known, rigid object and its subsequent grasping is to be
performed, remains nearly unchanging and might even be known to the robot in
advance. In this paper, we refer to this problem as instance-specific pose
estimation: the robot is expected to estimate the pose with a high degree of
accuracy in only a limited set of familiar scenarios. Minor changes in the
scene, including variations in lighting conditions and background appearance,
are acceptable but drastic alterations are not anticipated. To this end, we
present a method to rapidly train and deploy a pipeline for estimating the
continuous 6-DoF pose of an object from a single RGB image. The key idea is to
leverage known camera poses and rigid body geometry to partially automate the
generation of a large labeled dataset. The dataset, along with sufficient
domain randomization, is then used to supervise the training of deep neural
networks for predicting semantic keypoints. Experimentally, we demonstrate the
convenience and effectiveness of our proposed method to accurately estimate
object pose requiring only a very small amount of manual annotation for
training.
Related papers
- Category Level 6D Object Pose Estimation from a Single RGB Image using Diffusion [9.025235713063509]
We tackle the harder problem of pose estimation for category-level objects from a single RGB image.
We propose a novel solution that eliminates the need for specific object models or depth information.
Our approach outperforms the current state-of-the-art on the REAL275 dataset by a significant margin.
arXiv Detail & Related papers (2024-12-16T03:39:33Z) - YOLOPose V2: Understanding and Improving Transformer-based 6D Pose
Estimation [36.067414358144816]
YOLOPose is a Transformer-based multi-object 6D pose estimation method.
We employ a learnable orientation estimation module to predict the orientation from the keypoints.
Our method is suitable for real-time applications and achieves results comparable to state-of-the-art methods.
arXiv Detail & Related papers (2023-07-21T12:53:54Z) - Learning to Estimate 6DoF Pose from Limited Data: A Few-Shot,
Generalizable Approach using RGB Images [60.0898989456276]
We present a new framework named Cas6D for few-shot 6DoF pose estimation that is generalizable and uses only RGB images.
To address the false positives of target object detection in the extreme few-shot setting, our framework utilizes a self-supervised pre-trained ViT to learn robust feature representations.
Experimental results on the LINEMOD and GenMOP datasets demonstrate that Cas6D outperforms state-of-the-art methods by 9.2% and 3.8% accuracy (Proj-5) under the 32-shot setting.
arXiv Detail & Related papers (2023-06-13T07:45:42Z) - Ambiguity-Aware Multi-Object Pose Optimization for Visually-Assisted
Robot Manipulation [17.440729138126162]
We present an ambiguity-aware 6D object pose estimation network, PrimA6D++, as a generic uncertainty prediction method.
The proposed method shows a significant performance improvement in T-LESS and YCB-Video datasets.
We further demonstrate real-time scene recognition capability for visually-assisted robot manipulation.
arXiv Detail & Related papers (2022-11-02T08:57:20Z) - Unseen Object 6D Pose Estimation: A Benchmark and Baselines [62.8809734237213]
We propose a new task that enables and facilitates algorithms to estimate the 6D pose estimation of novel objects during testing.
We collect a dataset with both real and synthetic images and up to 48 unseen objects in the test set.
By training an end-to-end 3D correspondences network, our method finds corresponding points between an unseen object and a partial view RGBD image accurately and efficiently.
arXiv Detail & Related papers (2022-06-23T16:29:53Z) - FS6D: Few-Shot 6D Pose Estimation of Novel Objects [116.34922994123973]
6D object pose estimation networks are limited in their capability to scale to large numbers of object instances.
In this work, we study a new open set problem; the few-shot 6D object poses estimation: estimating the 6D pose of an unknown object by a few support views without extra training.
arXiv Detail & Related papers (2022-03-28T10:31:29Z) - Spatial Attention Improves Iterative 6D Object Pose Estimation [52.365075652976735]
We propose a new method for 6D pose estimation refinement from RGB images.
Our main insight is that after the initial pose estimate, it is important to pay attention to distinct spatial features of the object.
We experimentally show that this approach learns to attend to salient spatial features and learns to ignore occluded parts of the object, leading to better pose estimation across datasets.
arXiv Detail & Related papers (2021-01-05T17:18:52Z) - Continuous close-range 3D object pose estimation [1.4502611532302039]
Vision-based 3D pose estimation is a necessity to accurately handle objects that might not be placed at fixed positions.
In this paper, we present a 3D pose estimation method based on a gradient-ascend particle filter.
Thereby, we can apply this method online during task execution to save valuable cycle time.
arXiv Detail & Related papers (2020-10-02T07:48:17Z) - I Like to Move It: 6D Pose Estimation as an Action Decision Process [53.63776807432945]
Object pose estimation is an integral part of robot vision and AR.
Previous 6D pose retrieval pipelines treat the problem either as a regression task or discretize the pose space to classify.
We change this paradigm and reformulate the problem as an action decision process where an initial pose is updated in incremental discrete steps.
A neural network estimates likely moves from a single RGB image iteratively and determines so an acceptable final pose.
arXiv Detail & Related papers (2020-09-26T20:05:42Z) - CPS++: Improving Class-level 6D Pose and Shape Estimation From Monocular
Images With Self-Supervised Learning [74.53664270194643]
Modern monocular 6D pose estimation methods can only cope with a handful of object instances.
We propose a novel method for class-level monocular 6D pose estimation, coupled with metric shape retrieval.
We experimentally demonstrate that we can retrieve precise 6D poses and metric shapes from a single RGB image.
arXiv Detail & Related papers (2020-03-12T15:28:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.