Semantic State Estimation in Cloth Manipulation Tasks
- URL: http://arxiv.org/abs/2203.11647v1
- Date: Tue, 22 Mar 2022 11:59:52 GMT
- Title: Semantic State Estimation in Cloth Manipulation Tasks
- Authors: Georgies Tzelepis, Eren Erdal Aksoy, J\'ulia Borr\`as, and Guillem
Aleny\`a
- Abstract summary: In this paper, we aim to solve the problem of semantic state estimation in cloth manipulation tasks.
We introduce a new large-scale fully-annotated RGB image dataset showing various human demonstrations of different complicated cloth manipulations.
We provide a set of baseline deep networks and benchmark them on the problem of semantic state estimation.
- Score: 0.4812321790984493
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Understanding of deformable object manipulations such as textiles is a
challenge due to the complexity and high dimensionality of the problem.
Particularly, the lack of a generic representation of semantic states (e.g.,
\textit{crumpled}, \textit{diagonally folded}) during a continuous manipulation
process introduces an obstacle to identify the manipulation type. In this
paper, we aim to solve the problem of semantic state estimation in cloth
manipulation tasks. For this purpose, we introduce a new large-scale
fully-annotated RGB image dataset showing various human demonstrations of
different complicated cloth manipulations. We provide a set of baseline deep
networks and benchmark them on the problem of semantic state estimation using
our proposed dataset. Furthermore, we investigate the scalability of our
semantic state estimation framework in robot monitoring tasks of long and
complex cloth manipulations.
Related papers
- Towards Confidence-guided Shape Completion for Robotic Applications [6.940242990198]
Deep learning has begun taking traction as effective means of inferring a complete 3D object representation from partial visual data.
We propose an object shape completion method based on an implicit 3D representation providing a confidence value for each reconstructed point.
We experimentally validate our approach by comparing reconstructed shapes with ground truths, and by deploying our shape completion algorithm in a robotic grasping pipeline.
arXiv Detail & Related papers (2022-09-09T13:48:24Z) - MetaGraspNet: A Large-Scale Benchmark Dataset for Scene-Aware
Ambidextrous Bin Picking via Physics-based Metaverse Synthesis [72.85526892440251]
We introduce MetaGraspNet, a large-scale photo-realistic bin picking dataset constructed via physics-based metaverse synthesis.
The proposed dataset contains 217k RGBD images across 82 different article types, with full annotations for object detection, amodal perception, keypoint detection, manipulation order and ambidextrous grasp labels for a parallel-jaw and vacuum gripper.
We also provide a real dataset consisting of over 2.3k fully annotated high-quality RGBD images, divided into 5 levels of difficulties and an unseen object set to evaluate different object and layout properties.
arXiv Detail & Related papers (2022-08-08T08:15:34Z) - An End-to-End Approach for Seam Carving Detection using Deep Neural
Networks [0.0]
Seam carving is a computational method capable of resizing images for both reduction and expansion based on its content, instead of the image geometry.
We propose an end-to-end approach to cope with the problem of automatic seam carving detection.
arXiv Detail & Related papers (2022-03-05T12:53:55Z) - Object-aware Monocular Depth Prediction with Instance Convolutions [72.98771405534937]
We propose a novel convolutional operator which is explicitly tailored to avoid feature aggregation.
Our method is based on estimating per-part depth values by means of superpixels.
Our evaluation with respect to the NYUv2 as well as the iBims dataset clearly demonstrates the superiority of Instance Convolutions.
arXiv Detail & Related papers (2021-12-02T18:59:48Z) - RICE: Refining Instance Masks in Cluttered Environments with Graph
Neural Networks [53.15260967235835]
We propose a novel framework that refines the output of such methods by utilizing a graph-based representation of instance masks.
We train deep networks capable of sampling smart perturbations to the segmentations, and a graph neural network, which can encode relations between objects, to evaluate the segmentations.
We demonstrate an application that uses uncertainty estimates generated by our method to guide a manipulator, leading to efficient understanding of cluttered scenes.
arXiv Detail & Related papers (2021-06-29T20:29:29Z) - Differentiable Multi-Granularity Human Representation Learning for
Instance-Aware Human Semantic Parsing [131.97475877877608]
A new bottom-up regime is proposed to learn category-level human semantic segmentation and multi-person pose estimation in a joint and end-to-end manner.
It is a compact, efficient and powerful framework that exploits structural information over different human granularities.
Experiments on three instance-aware human datasets show that our model outperforms other bottom-up alternatives with much more efficient inference.
arXiv Detail & Related papers (2021-03-08T06:55:00Z) - Video Anomaly Detection by Estimating Likelihood of Representations [21.879366166261228]
Video anomaly is a challenging task because it involves solving many sub-tasks such as motion representation, object localization and action recognition.
Traditionally, solutions to this task have focused on the mapping between video frames and their low-dimensional features, while ignoring the spatial connections of those features.
Recent solutions focus on analyzing these spatial connections by using hard clustering techniques, such as K-Means, or applying neural networks to map latent features to a general understanding.
In order to solve video anomaly in the latent feature space, we propose a deep probabilistic model to transfer this task into a density estimation problem.
arXiv Detail & Related papers (2020-12-02T19:16:22Z) - Encoding cloth manipulations using a graph of states and transitions [8.778914180886835]
We propose a generic, compact and simplified representation of the states of cloth manipulation.
We also define a Cloth Manipulation Graph that encodes all the strategies to accomplish a task.
arXiv Detail & Related papers (2020-09-30T13:56:13Z) - Adversarial Semantic Data Augmentation for Human Pose Estimation [96.75411357541438]
We propose Semantic Data Augmentation (SDA), a method that augments images by pasting segmented body parts with various semantic granularity.
We also propose Adversarial Semantic Data Augmentation (ASDA), which exploits a generative network to dynamiclly predict tailored pasting configuration.
State-of-the-art results are achieved on challenging benchmarks.
arXiv Detail & Related papers (2020-08-03T07:56:04Z) - Efficient State Abstraction using Object-centered Predicates for
Manipulation Planning [86.24148040040885]
We propose an object-centered representation that permits characterizing a much wider set of possible changes in configuration spaces.
Based on this representation, we define universal planning operators for picking and placing actions that permits generating plans with geometric and force consistency.
arXiv Detail & Related papers (2020-07-16T10:52:53Z) - Robust, Occlusion-aware Pose Estimation for Objects Grasped by Adaptive
Hands [16.343365158924183]
manipulation tasks, such as within-hand manipulation, require the object's pose relative to a robot hand.
This paper presents a depth-based framework, which aims for robust pose estimation and short response times.
arXiv Detail & Related papers (2020-03-07T05:51:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.