Unseen Object Instance Segmentation for Robotic Environments
- URL: http://arxiv.org/abs/2007.08073v2
- Date: Thu, 14 Oct 2021 02:56:33 GMT
- Title: Unseen Object Instance Segmentation for Robotic Environments
- Authors: Christopher Xie, Yu Xiang, Arsalan Mousavian, Dieter Fox
- Abstract summary: We propose a method to segment unseen object instances in tabletop environments.
UOIS-Net is comprised of two stages: first, it operates only on depth to produce object instance center votes in 2D or 3D.
Surprisingly, our framework is able to learn from synthetic RGB-D data where the RGB is non-photorealistic.
- Score: 67.88276573341734
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In order to function in unstructured environments, robots need the ability to
recognize unseen objects. We take a step in this direction by tackling the
problem of segmenting unseen object instances in tabletop environments.
However, the type of large-scale real-world dataset required for this task
typically does not exist for most robotic settings, which motivates the use of
synthetic data. Our proposed method, UOIS-Net, separately leverages synthetic
RGB and synthetic depth for unseen object instance segmentation. UOIS-Net is
comprised of two stages: first, it operates only on depth to produce object
instance center votes in 2D or 3D and assembles them into rough initial masks.
Secondly, these initial masks are refined using RGB. Surprisingly, our
framework is able to learn from synthetic RGB-D data where the RGB is
non-photorealistic. To train our method, we introduce a large-scale synthetic
dataset of random objects on tabletops. We show that our method can produce
sharp and accurate segmentation masks, outperforming state-of-the-art methods
on unseen object instance segmentation. We also show that our method can
segment unseen objects for robot grasping.
Related papers
- LAC-Net: Linear-Fusion Attention-Guided Convolutional Network for Accurate Robotic Grasping Under the Occlusion [79.22197702626542]
This paper introduces a framework that explores amodal segmentation for robotic grasping in cluttered scenes.
We propose a Linear-fusion Attention-guided Convolutional Network (LAC-Net)
The results on different datasets show that our method achieves state-of-the-art performance.
arXiv Detail & Related papers (2024-08-06T14:50:48Z) - Learning Sim-to-Real Dense Object Descriptors for Robotic Manipulation [4.7246285569677315]
We present Sim-to-Real Dense Object Nets (SRDONs), a dense object descriptor that not only understands the object via appropriate representation but also maps simulated and real data to a unified feature space with pixel consistency.
We demonstrate in experiments that pre-trained SRDONs significantly improve performances on unseen objects and unseen visual environments for various robotic tasks with zero real-world training.
arXiv Detail & Related papers (2023-04-18T02:28:55Z) - SupeRGB-D: Zero-shot Instance Segmentation in Cluttered Indoor
Environments [67.34330257205525]
In this work, we explore zero-shot instance segmentation (ZSIS) from RGB-D data to identify unseen objects in a semantic category-agnostic manner.
We present a method that uses annotated objects to learn the objectness'' of pixels and generalize to unseen object categories in cluttered indoor environments.
arXiv Detail & Related papers (2022-12-22T17:59:48Z) - Unseen Object Instance Segmentation with Fully Test-time RGB-D
Embeddings Adaptation [14.258456366985444]
Recently, a popular solution is leveraging RGB-D features of large-scale synthetic data and applying the model to unseen real-world scenarios.
We re-emphasize the adaptation process across Sim2Real domains in this paper.
We propose a framework to conduct the Fully Test-time RGB-D Embeddings Adaptation (FTEA) based on parameters of the BatchNorm layer.
arXiv Detail & Related papers (2022-04-21T02:35:20Z) - Discovering Objects that Can Move [55.743225595012966]
We study the problem of object discovery -- separating objects from the background without manual labels.
Existing approaches utilize appearance cues, such as color, texture, and location, to group pixels into object-like regions.
We choose to focus on dynamic objects -- entities that can move independently in the world.
arXiv Detail & Related papers (2022-03-18T21:13:56Z) - IFOR: Iterative Flow Minimization for Robotic Object Rearrangement [92.97142696891727]
IFOR, Iterative Flow Minimization for Robotic Object Rearrangement, is an end-to-end method for the problem of object rearrangement for unknown objects.
We show that our method applies to cluttered scenes, and in the real world, while training only on synthetic data.
arXiv Detail & Related papers (2022-02-01T20:03:56Z) - Unseen Object Amodal Instance Segmentation via Hierarchical Occlusion
Modeling [0.0]
Instance-aware segmentation of unseen objects is essential for a robotic system in an unstructured environment.
This paper addresses Unseen Object Amodal Instances (UOAIS) to detect 1) visible masks, 2) amodal masks, and 3) occlusions on unseen object instances.
We evaluate our method on three benchmarks (tabletop, indoors, and bin environments) and achieved state-of-the-art (SOTA) performance.
arXiv Detail & Related papers (2021-09-23T01:55:42Z) - Learning RGB-D Feature Embeddings for Unseen Object Instance
Segmentation [67.88276573341734]
We propose a new method for unseen object instance segmentation by learning RGB-D feature embeddings from synthetic data.
A metric learning loss function is utilized to learn to produce pixel-wise feature embeddings.
We further improve the segmentation accuracy with a new two-stage clustering algorithm.
arXiv Detail & Related papers (2020-07-30T00:23:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.