YCB-M: A Multi-Camera RGB-D Dataset for Object Recognition and 6DoF Pose
Estimation
- URL: http://arxiv.org/abs/2004.11657v2
- Date: Tue, 29 Sep 2020 07:58:18 GMT
- Title: YCB-M: A Multi-Camera RGB-D Dataset for Object Recognition and 6DoF Pose
Estimation
- Authors: Till Grenzd\"orffer, Martin G\"unther and Joachim Hertzberg
- Abstract summary: We present a dataset of 32 scenes that have been captured by 7 different 3D cameras, totaling 49,294 frames.
This allows evaluating the sensitivity of pose estimation algorithms to the specifics of the used camera.
- Score: 2.9972063833424216
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: While a great variety of 3D cameras have been introduced in recent years,
most publicly available datasets for object recognition and pose estimation
focus on one single camera. In this work, we present a dataset of 32 scenes
that have been captured by 7 different 3D cameras, totaling 49,294 frames. This
allows evaluating the sensitivity of pose estimation algorithms to the
specifics of the used camera and the development of more robust algorithms that
are more independent of the camera model. Vice versa, our dataset enables
researchers to perform a quantitative comparison of the data from several
different cameras and depth sensing technologies and evaluate their algorithms
before selecting a camera for their specific task. The scenes in our dataset
contain 20 different objects from the common benchmark YCB object and model set
[1], [2]. We provide full ground truth 6DoF poses for each object, per-pixel
segmentation, 2D and 3D bounding boxes and a measure of the amount of occlusion
of each object. We have also performed an initial evaluation of the cameras
using our dataset on a state-of-the-art object recognition and pose estimation
system [3].
Related papers
- LocaliseBot: Multi-view 3D object localisation with differentiable
rendering for robot grasping [9.690844449175948]
We focus on object pose estimation.
Our approach relies on three pieces of information: multiple views of the object, the camera's parameters at those viewpoints, and 3D CAD models of objects.
We show that the estimated object pose results in 99.65% grasp accuracy with the ground truth grasp candidates.
arXiv Detail & Related papers (2023-11-14T14:27:53Z) - Unseen Object 6D Pose Estimation: A Benchmark and Baselines [62.8809734237213]
We propose a new task that enables and facilitates algorithms to estimate the 6D pose estimation of novel objects during testing.
We collect a dataset with both real and synthetic images and up to 48 unseen objects in the test set.
By training an end-to-end 3D correspondences network, our method finds corresponding points between an unseen object and a partial view RGBD image accurately and efficiently.
arXiv Detail & Related papers (2022-06-23T16:29:53Z) - Coupled Iterative Refinement for 6D Multi-Object Pose Estimation [64.7198752089041]
Given a set of known 3D objects and an RGB or RGB-D input image, we detect and estimate the 6D pose of each object.
Our approach iteratively refines both pose and correspondence in a tightly coupled manner, allowing us to dynamically remove outliers to improve accuracy.
arXiv Detail & Related papers (2022-04-26T18:00:08Z) - A Critical Analysis of Image-based Camera Pose Estimation Techniques [18.566761146552537]
Camera localization could benefit many computer vision fields, such as autonomous driving, robot navigation, and augmented reality (AR)
In this survey, we first introduce specific application areas and the evaluation metrics for camera localization pose according to different sub-tasks.
arXiv Detail & Related papers (2022-01-15T09:57:45Z) - MonoCInIS: Camera Independent Monocular 3D Object Detection using
Instance Segmentation [55.96577490779591]
Methods need to have a degree of 'camera independence' in order to benefit from large and heterogeneous training data.
We show that more data does not automatically guarantee a better performance, but rather, methods need to have a degree of 'camera independence' in order to benefit from large and heterogeneous training data.
arXiv Detail & Related papers (2021-10-01T14:56:37Z) - VoxelTrack: Multi-Person 3D Human Pose Estimation and Tracking in the
Wild [98.69191256693703]
We present VoxelTrack for multi-person 3D pose estimation and tracking from a few cameras which are separated by wide baselines.
It employs a multi-branch network to jointly estimate 3D poses and re-identification (Re-ID) features for all people in the environment.
It outperforms the state-of-the-art methods by a large margin on three public datasets including Shelf, Campus and CMU Panoptic.
arXiv Detail & Related papers (2021-08-05T08:35:44Z) - CosyPose: Consistent multi-view multi-object 6D pose estimation [48.097599674329004]
We present a single-view single-object 6D pose estimation method, which we use to generate 6D object pose hypotheses.
Second, we develop a robust method for matching individual 6D object pose hypotheses across different input images.
Third, we develop a method for global scene refinement given multiple object hypotheses and their correspondences across views.
arXiv Detail & Related papers (2020-08-19T14:11:56Z) - SIDOD: A Synthetic Image Dataset for 3D Object Pose Recognition with
Distractors [10.546457120988494]
This dataset contains 144k stereo image pairs that synthetically combine 18 camera viewpoints of three photorealistic virtual environments with up to 10 objects.
We describe our approach for domain randomization and provide insight into the decisions that produced the dataset.
arXiv Detail & Related papers (2020-08-12T00:14:19Z) - Single View Metrology in the Wild [94.7005246862618]
We present a novel approach to single view metrology that can recover the absolute scale of a scene represented by 3D heights of objects or camera height above the ground.
Our method relies on data-driven priors learned by a deep network specifically designed to imbibe weakly supervised constraints from the interplay of the unknown camera with 3D entities such as object heights.
We demonstrate state-of-the-art qualitative and quantitative results on several datasets as well as applications including virtual object insertion.
arXiv Detail & Related papers (2020-07-18T22:31:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.