Instance Segmentation of Visible and Occluded Regions for Finding and
Picking Target from a Pile of Objects
- URL: http://arxiv.org/abs/2001.07475v1
- Date: Tue, 21 Jan 2020 12:28:37 GMT
- Title: Instance Segmentation of Visible and Occluded Regions for Finding and
Picking Target from a Pile of Objects
- Authors: Kentaro Wada, Shingo Kitagawa, Kei Okada, Masayuki Inaba
- Abstract summary: We present a robotic system for picking a target from a pile of objects that is capable of finding and grasping the target object.
We extend an existing instance segmentation model with a novel relook' architecture, in which the model explicitly learns the inter-instance relationship.
Also, by using image synthesis, we make the system capable of handling new objects without human annotations.
- Score: 25.836334764387498
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present a robotic system for picking a target from a pile of objects that
is capable of finding and grasping the target object by removing obstacles in
the appropriate order. The fundamental idea is to segment instances with both
visible and occluded masks, which we call `instance occlusion segmentation'. To
achieve this, we extend an existing instance segmentation model with a novel
`relook' architecture, in which the model explicitly learns the inter-instance
relationship. Also, by using image synthesis, we make the system capable of
handling new objects without human annotations. The experimental results show
the effectiveness of the relook architecture when compared with a conventional
model and of the image synthesis when compared to a human-annotated dataset. We
also demonstrate the capability of our system to achieve picking a target in a
cluttered environment with a real robot.
Related papers
- ICGNet: A Unified Approach for Instance-Centric Grasping [42.92991092305974]
We introduce an end-to-end architecture for object-centric grasping.
We show the effectiveness of the proposed method by extensively evaluating it against state-of-the-art methods on synthetic datasets.
arXiv Detail & Related papers (2024-01-18T12:41:41Z) - Weakly-supervised Contrastive Learning for Unsupervised Object Discovery [52.696041556640516]
Unsupervised object discovery is promising due to its ability to discover objects in a generic manner.
We design a semantic-guided self-supervised learning model to extract high-level semantic features from images.
We introduce Principal Component Analysis (PCA) to localize object regions.
arXiv Detail & Related papers (2023-07-07T04:03:48Z) - Perceive, Excavate and Purify: A Novel Object Mining Framework for
Instance Segmentation [4.375012768093524]
We propose a novel object mining framework for instance segmentation.
We first introduce the semantics perceiving subnetwork to capture pixels that may belong to an obvious instance from the bottom up.
In the mechanism, preliminary perceived semantics are regarded as original instances with classifications and locations, and then indistinguishable objects around these original instances are mined.
arXiv Detail & Related papers (2023-04-18T08:47:03Z) - Complex-Valued Autoencoders for Object Discovery [62.26260974933819]
We propose a distributed approach to object-centric representations: the Complex AutoEncoder.
We show that this simple and efficient approach achieves better reconstruction performance than an equivalent real-valued autoencoder on simple multi-object datasets.
We also show that it achieves competitive unsupervised object discovery performance to a SlotAttention model on two datasets, and manages to disentangle objects in a third dataset where SlotAttention fails - all while being 7-70 times faster to train.
arXiv Detail & Related papers (2022-04-05T09:25:28Z) - Discovering Objects that Can Move [55.743225595012966]
We study the problem of object discovery -- separating objects from the background without manual labels.
Existing approaches utilize appearance cues, such as color, texture, and location, to group pixels into object-like regions.
We choose to focus on dynamic objects -- entities that can move independently in the world.
arXiv Detail & Related papers (2022-03-18T21:13:56Z) - Semantically Grounded Object Matching for Robust Robotic Scene
Rearrangement [21.736603698556042]
We present a novel approach to object matching that uses a large pre-trained vision-language model to match objects in a cross-instance setting.
We demonstrate that this provides considerably improved matching performance in cross-instance settings.
arXiv Detail & Related papers (2021-11-15T18:39:43Z) - Unseen Object Amodal Instance Segmentation via Hierarchical Occlusion
Modeling [0.0]
Instance-aware segmentation of unseen objects is essential for a robotic system in an unstructured environment.
This paper addresses Unseen Object Amodal Instances (UOAIS) to detect 1) visible masks, 2) amodal masks, and 3) occlusions on unseen object instances.
We evaluate our method on three benchmarks (tabletop, indoors, and bin environments) and achieved state-of-the-art (SOTA) performance.
arXiv Detail & Related papers (2021-09-23T01:55:42Z) - RICE: Refining Instance Masks in Cluttered Environments with Graph
Neural Networks [53.15260967235835]
We propose a novel framework that refines the output of such methods by utilizing a graph-based representation of instance masks.
We train deep networks capable of sampling smart perturbations to the segmentations, and a graph neural network, which can encode relations between objects, to evaluate the segmentations.
We demonstrate an application that uses uncertainty estimates generated by our method to guide a manipulator, leading to efficient understanding of cluttered scenes.
arXiv Detail & Related papers (2021-06-29T20:29:29Z) - Synthesizing the Unseen for Zero-shot Object Detection [72.38031440014463]
We propose to synthesize visual features for unseen classes, so that the model learns both seen and unseen objects in the visual domain.
We use a novel generative model that uses class-semantics to not only generate the features but also to discriminatively separate them.
arXiv Detail & Related papers (2020-10-19T12:36:11Z) - DyStaB: Unsupervised Object Segmentation via Dynamic-Static
Bootstrapping [72.84991726271024]
We describe an unsupervised method to detect and segment portions of images of live scenes that are seen moving as a coherent whole.
Our method first partitions the motion field by minimizing the mutual information between segments.
It uses the segments to learn object models that can be used for detection in a static image.
arXiv Detail & Related papers (2020-08-16T22:05:13Z) - A Deep Learning Approach to Object Affordance Segmentation [31.221897360610114]
We design an autoencoder that infers pixel-wise affordance labels in both videos and static images.
Our model surpasses the need for object labels and bounding boxes by using a soft-attention mechanism.
We show that our model achieves competitive results compared to strongly supervised methods on SOR3D-AFF.
arXiv Detail & Related papers (2020-04-18T15:34:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.