BOP Challenge 2022 on Detection, Segmentation and Pose Estimation of
Specific Rigid Objects
- URL: http://arxiv.org/abs/2302.13075v1
- Date: Sat, 25 Feb 2023 13:12:50 GMT
- Title: BOP Challenge 2022 on Detection, Segmentation and Pose Estimation of
Specific Rigid Objects
- Authors: Martin Sundermeyer, Tomas Hodan, Yann Labbe, Gu Wang, Eric Brachmann,
Bertram Drost, Carsten Rother, Jiri Matas
- Abstract summary: Methods based on point pair features, which were introduced in 2010, are now clearly outperformed by deep learning methods.
The synthetic-to-real domain gap was again significantly reduced.
The fastest variant of GDRNPP reached 80.5 AR$_C$ with an average time per image of 0.23s.
- Score: 59.70444717956975
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We present the evaluation methodology, datasets and results of the BOP
Challenge 2022, the fourth in a series of public competitions organized with
the goal to capture the status quo in the field of 6D object pose estimation
from an RGB/RGB-D image. In 2022, we witnessed another significant improvement
in the pose estimation accuracy -- the state of the art, which was 56.9 AR$_C$
in 2019 (Vidal et al.) and 69.8 AR$_C$ in 2020 (CosyPose), moved to new heights
of 83.7 AR$_C$ (GDRNPP). Out of 49 pose estimation methods evaluated since
2019, the top 18 are from 2022. Methods based on point pair features, which
were introduced in 2010 and achieved competitive results even in 2020, are now
clearly outperformed by deep learning methods. The synthetic-to-real domain gap
was again significantly reduced, with 82.7 AR$_C$ achieved by GDRNPP trained
only on synthetic images from BlenderProc. The fastest variant of GDRNPP
reached 80.5 AR$_C$ with an average time per image of 0.23s. Since most of the
recent methods for 6D object pose estimation begin by detecting/segmenting
objects, we also started evaluating 2D object detection and segmentation
performance based on the COCO metrics. Compared to the Mask R-CNN results from
CosyPose in 2020, detection improved from 60.3 to 77.3 AP$_C$ and segmentation
from 40.5 to 58.7 AP$_C$. The online evaluation system stays open and is
available at: \href{http://bop.felk.cvut.cz/}{bop.felk.cvut.cz}.
Related papers
- BOP Challenge 2023 on Detection, Segmentation and Pose Estimation of Seen and Unseen Rigid Objects [54.90773237124648]
The BOP Challenge 2023 is the fifth in a series of public competitions organized to capture the state of the art in model-based 6D object pose estimation.
The best method for seen objects (GPose) achieved a moderate accuracy improvement but a significant 43% run-time improvement compared to the best 2022 counterpart.
arXiv Detail & Related papers (2024-03-14T18:37:46Z) - Learning to Estimate 6DoF Pose from Limited Data: A Few-Shot,
Generalizable Approach using RGB Images [60.0898989456276]
We present a new framework named Cas6D for few-shot 6DoF pose estimation that is generalizable and uses only RGB images.
To address the false positives of target object detection in the extreme few-shot setting, our framework utilizes a self-supervised pre-trained ViT to learn robust feature representations.
Experimental results on the LINEMOD and GenMOP datasets demonstrate that Cas6D outperforms state-of-the-art methods by 9.2% and 3.8% accuracy (Proj-5) under the 32-shot setting.
arXiv Detail & Related papers (2023-06-13T07:45:42Z) - Unseen Object 6D Pose Estimation: A Benchmark and Baselines [62.8809734237213]
We propose a new task that enables and facilitates algorithms to estimate the 6D pose estimation of novel objects during testing.
We collect a dataset with both real and synthetic images and up to 48 unseen objects in the test set.
By training an end-to-end 3D correspondences network, our method finds corresponding points between an unseen object and a partial view RGBD image accurately and efficiently.
arXiv Detail & Related papers (2022-06-23T16:29:53Z) - Category-Level 6D Object Pose Estimation via Cascaded Relation and
Recurrent Reconstruction Networks [22.627704070200863]
Category-level 6D pose estimation is fundamental to many scenarios such as robotic manipulation and augmented reality.
We achieve accurate category-level 6D pose estimation via cascaded relation and recurrent reconstruction networks.
Our method exceeds the latest state-of-the-art SPD by $4.9%$ and $17.7%$ on the CAMERA25 dataset.
arXiv Detail & Related papers (2021-08-19T15:46:52Z) - Sparse R-CNN: End-to-End Object Detection with Learnable Proposals [77.9701193170127]
We present Sparse R-CNN, a purely sparse method for object detection in images.
Final predictions are directly output without non-maximum suppression post-procedure.
We hope our work could inspire re-thinking the convention of dense prior in object detectors.
arXiv Detail & Related papers (2020-11-25T00:01:28Z) - EfficientPose: An efficient, accurate and scalable end-to-end 6D multi
object pose estimation approach [0.0]
We introduce EfficientPose, a new approach for 6D object pose estimation.
It is highly accurate, efficient and scalable over a wide range of computational resources.
It can detect the 2D bounding box of multiple objects and instances as well as estimate their full 6D poses in a single shot.
arXiv Detail & Related papers (2020-11-09T10:23:55Z) - Learning Delicate Local Representations for Multi-Person Pose Estimation [77.53144055780423]
We propose a novel method called Residual Steps Network (RSN)
RSN aggregates features with the same spatial size (Intra-level features) efficiently to obtain delicate local representations.
Our approach won the 1st place of COCO Keypoint Challenge 2019.
arXiv Detail & Related papers (2020-03-09T10:40:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.