Model-Based Underwater 6D Pose Estimation from RGB
- URL: http://arxiv.org/abs/2302.06821v2
- Date: Fri, 15 Sep 2023 09:34:12 GMT
- Title: Model-Based Underwater 6D Pose Estimation from RGB
- Authors: Davide Sapienza, Elena Govi, Sara Aldhaheri, Marko Bertogna, Eloy
Roura, \`Eric Pairet, Micaela Verucchi, Paola Ard\'on
- Abstract summary: We propose an approach that leverages 2D object detection to reliably compute 6D pose estimates in different underwater scenarios.
All objects and scenes are made available in an open-source dataset that includes annotations for object detection and pose estimation.
- Score: 1.9160624126555885
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Object pose estimation underwater allows an autonomous system to perform
tracking and intervention tasks. Nonetheless, underwater target pose estimation
is remarkably challenging due to, among many factors, limited visibility, light
scattering, cluttered environments, and constantly varying water conditions. An
approach is to employ sonar or laser sensing to acquire 3D data, however, the
data is not clear and the sensors expensive. For this reason, the community has
focused on extracting pose estimates from RGB input. In this work, we propose
an approach that leverages 2D object detection to reliably compute 6D pose
estimates in different underwater scenarios. We test our proposal with 4
objects with symmetrical shapes and poor texture spanning across 33,920
synthetic and 10 real scenes. All objects and scenes are made available in an
open-source dataset that includes annotations for object detection and pose
estimation. When benchmarking against similar end-to-end methodologies for 6D
object pose estimation, our pipeline provides estimates that are 8% more
accurate. We also demonstrate the real world usability of our pose estimation
pipeline on an underwater robotic manipulator in a reaching task.
Related papers
- FAFA: Frequency-Aware Flow-Aided Self-Supervision for Underwater Object Pose Estimation [65.01601309903971]
We introduce FAFA, a Frequency-Aware Flow-Aided self-supervised framework for 6D pose estimation of unmanned underwater vehicles (UUVs)
Our framework relies solely on the 3D model and RGB images, alleviating the need for any real pose annotations or other-modality data like depths.
We evaluate the effectiveness of FAFA on common underwater object pose benchmarks and showcase significant performance improvements compared to state-of-the-art methods.
arXiv Detail & Related papers (2024-09-25T03:54:01Z) - LocaliseBot: Multi-view 3D object localisation with differentiable
rendering for robot grasping [9.690844449175948]
We focus on object pose estimation.
Our approach relies on three pieces of information: multiple views of the object, the camera's parameters at those viewpoints, and 3D CAD models of objects.
We show that the estimated object pose results in 99.65% grasp accuracy with the ground truth grasp candidates.
arXiv Detail & Related papers (2023-11-14T14:27:53Z) - Semi-Perspective Decoupled Heatmaps for 3D Robot Pose Estimation from
Depth Maps [66.24554680709417]
Knowing the exact 3D location of workers and robots in a collaborative environment enables several real applications.
We propose a non-invasive framework based on depth devices and deep neural networks to estimate the 3D pose of robots from an external camera.
arXiv Detail & Related papers (2022-07-06T08:52:12Z) - Unseen Object 6D Pose Estimation: A Benchmark and Baselines [62.8809734237213]
We propose a new task that enables and facilitates algorithms to estimate the 6D pose estimation of novel objects during testing.
We collect a dataset with both real and synthetic images and up to 48 unseen objects in the test set.
By training an end-to-end 3D correspondences network, our method finds corresponding points between an unseen object and a partial view RGBD image accurately and efficiently.
arXiv Detail & Related papers (2022-06-23T16:29:53Z) - Object detection and Autoencoder-based 6D pose estimation for highly
cluttered Bin Picking [14.076644545879939]
We propose a framework for pose estimation in highly cluttered scenes with small objects.
In this work, we compare synthetic data generation approaches for object detection and pose estimation.
We introduce a pose filtering algorithm that determines the most accurate estimated poses.
arXiv Detail & Related papers (2021-06-15T11:01:07Z) - PLUME: Efficient 3D Object Detection from Stereo Images [95.31278688164646]
Existing methods tackle the problem in two steps: first depth estimation is performed, a pseudo LiDAR point cloud representation is computed from the depth estimates, and then object detection is performed in 3D space.
We propose a model that unifies these two tasks in the same metric space.
Our approach achieves state-of-the-art performance on the challenging KITTI benchmark, with significantly reduced inference time compared with existing methods.
arXiv Detail & Related papers (2021-01-17T05:11:38Z) - Spatial Attention Improves Iterative 6D Object Pose Estimation [52.365075652976735]
We propose a new method for 6D pose estimation refinement from RGB images.
Our main insight is that after the initial pose estimate, it is important to pay attention to distinct spatial features of the object.
We experimentally show that this approach learns to attend to salient spatial features and learns to ignore occluded parts of the object, leading to better pose estimation across datasets.
arXiv Detail & Related papers (2021-01-05T17:18:52Z) - Pose Estimation of Specular and Symmetrical Objects [0.719973338079758]
In the robotic industry, specular and textureless metallic components are ubiquitous.
The 6D pose estimation of such objects with only a monocular RGB camera is difficult because of the absence of rich texture features.
This paper proposes a data-driven solution to estimate the 6D pose of specular objects for grasping them.
arXiv Detail & Related papers (2020-10-31T22:08:46Z) - SHREC 2020 track: 6D Object Pose Estimation [26.4781238445338]
6D pose estimation is crucial for augmented reality, virtual reality, robotic manipulation and visual navigation.
Different pose estimation methods have different strengths and weaknesses, depending on feature representations and scene contents.
Existing 3D datasets that are used for data-driven methods to estimate 6D poses have limited view angles and low resolution.
arXiv Detail & Related papers (2020-10-19T09:45:42Z) - Measuring Generalisation to Unseen Viewpoints, Articulations, Shapes and
Objects for 3D Hand Pose Estimation under Hand-Object Interaction [137.28465645405655]
HANDS'19 is a challenge to evaluate the abilities of current 3D hand pose estimators (HPEs) to interpolate and extrapolate the poses of a training set.
We show that the accuracy of state-of-the-art methods can drop, and that they fail mostly on poses absent from the training set.
arXiv Detail & Related papers (2020-03-30T19:28:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.