Supervised Training of Dense Object Nets using Optimal Descriptors for
Industrial Robotic Applications
- URL: http://arxiv.org/abs/2102.08096v1
- Date: Tue, 16 Feb 2021 11:40:12 GMT
- Title: Supervised Training of Dense Object Nets using Optimal Descriptors for
Industrial Robotic Applications
- Authors: Andras Kupcsik, Markus Spies, Alexander Klein, Marco Todescato,
Nicolai Waniek, Philipp Schillinger, Mathias Buerger
- Abstract summary: Dense Object Nets (DONs) by Florence, Manuelli and Tedrake introduced dense object descriptors as a novel visual object representation for the robotics community.
In this paper we show that given a 3D model of an object, we can generate its descriptor space image, which allows for supervised training of DONs.
We compare the training methods on generating 6D grasps for industrial objects and show that our novel supervised training approach improves the pick-and-place performance in industry-relevant tasks.
- Score: 57.87136703404356
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Dense Object Nets (DONs) by Florence, Manuelli and Tedrake (2018) introduced
dense object descriptors as a novel visual object representation for the
robotics community. It is suitable for many applications including object
grasping, policy learning, etc. DONs map an RGB image depicting an object into
a descriptor space image, which implicitly encodes key features of an object
invariant to the relative camera pose. Impressively, the self-supervised
training of DONs can be applied to arbitrary objects and can be evaluated and
deployed within hours. However, the training approach relies on accurate depth
images and faces challenges with small, reflective objects, typical for
industrial settings, when using consumer grade depth cameras. In this paper we
show that given a 3D model of an object, we can generate its descriptor space
image, which allows for supervised training of DONs. We rely on Laplacian
Eigenmaps (LE) to embed the 3D model of an object into an optimally generated
space. While our approach uses more domain knowledge, it can be efficiently
applied even for smaller and reflective objects, as it does not rely on depth
information. We compare the training methods on generating 6D grasps for
industrial objects and show that our novel supervised training approach
improves the pick-and-place performance in industry-relevant tasks.
Related papers
- LocaliseBot: Multi-view 3D object localisation with differentiable
rendering for robot grasping [9.690844449175948]
We focus on object pose estimation.
Our approach relies on three pieces of information: multiple views of the object, the camera's parameters at those viewpoints, and 3D CAD models of objects.
We show that the estimated object pose results in 99.65% grasp accuracy with the ground truth grasp candidates.
arXiv Detail & Related papers (2023-11-14T14:27:53Z) - ROAM: Robust and Object-Aware Motion Generation Using Neural Pose
Descriptors [73.26004792375556]
This paper shows that robustness and generalisation to novel scene objects in 3D object-aware character synthesis can be achieved by training a motion model with as few as one reference object.
We leverage an implicit feature representation trained on object-only datasets, which encodes an SE(3)-equivariant descriptor field around the object.
We demonstrate substantial improvements in 3D virtual character motion and interaction quality and robustness to scenarios with unseen objects.
arXiv Detail & Related papers (2023-08-24T17:59:51Z) - MegaPose: 6D Pose Estimation of Novel Objects via Render & Compare [84.80956484848505]
MegaPose is a method to estimate the 6D pose of novel objects, that is, objects unseen during training.
We present a 6D pose refiner based on a render&compare strategy which can be applied to novel objects.
Second, we introduce a novel approach for coarse pose estimation which leverages a network trained to classify whether the pose error between a synthetic rendering and an observed image of the same object can be corrected by the refiner.
arXiv Detail & Related papers (2022-12-13T19:30:03Z) - DONet: Learning Category-Level 6D Object Pose and Size Estimation from
Depth Observation [53.55300278592281]
We propose a method of Category-level 6D Object Pose and Size Estimation (COPSE) from a single depth image.
Our framework makes inferences based on the rich geometric information of the object in the depth channel alone.
Our framework competes with state-of-the-art approaches that require labeled real-world images.
arXiv Detail & Related papers (2021-06-27T10:41:50Z) - Sparse Pose Trajectory Completion [87.31270669154452]
We propose a method to learn, even using a dataset where objects appear only in sparsely sampled views.
This is achieved with a cross-modal pose trajectory transfer mechanism.
Our method is evaluated on the Pix3D and ShapeNet datasets.
arXiv Detail & Related papers (2021-05-01T00:07:21Z) - Rapid Pose Label Generation through Sparse Representation of Unknown
Objects [7.32172860877574]
This work presents an approach for rapidly generating real-world, pose-annotated RGB-D data for unknown objects.
We first source minimalistic labelings of an ordered set of arbitrarily chosen keypoints over a set of RGB-D videos.
By solving an optimization problem, we combine these labels under a world frame to recover a sparse, keypoint-based representation of the object.
arXiv Detail & Related papers (2020-11-07T15:14:03Z) - Weakly Supervised Learning of Multi-Object 3D Scene Decompositions Using
Deep Shape Priors [69.02332607843569]
PriSMONet is a novel approach for learning Multi-Object 3D scene decomposition and representations from single images.
A recurrent encoder regresses a latent representation of 3D shape, pose and texture of each object from an input RGB image.
We evaluate the accuracy of our model in inferring 3D scene layout, demonstrate its generative capabilities, assess its generalization to real images, and point out benefits of the learned representation.
arXiv Detail & Related papers (2020-10-08T14:49:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.