Same Object, Different Grasps: Data and Semantic Knowledge for
Task-Oriented Grasping
- URL: http://arxiv.org/abs/2011.06431v2
- Date: Fri, 13 Nov 2020 15:28:44 GMT
- Title: Same Object, Different Grasps: Data and Semantic Knowledge for
Task-Oriented Grasping
- Authors: Adithyavairavan Murali, Weiyu Liu, Kenneth Marino, Sonia Chernova,
Abhinav Gupta
- Abstract summary: The TaskGrasp dataset is more diverse both in terms of objects and tasks, and an order of magnitude larger than previous datasets.
We present the GCNGrasp framework which uses the semantic knowledge of objects and tasks encoded in a knowledge graph to generalize to new object instances, classes and even new tasks.
We demonstrate that our dataset and model are applicable for the real world by executing task-oriented grasps on a real robot on unknown objects.
- Score: 40.95315009714416
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Despite the enormous progress and generalization in robotic grasping in
recent years, existing methods have yet to scale and generalize task-oriented
grasping to the same extent. This is largely due to the scale of the datasets
both in terms of the number of objects and tasks studied. We address these
concerns with the TaskGrasp dataset which is more diverse both in terms of
objects and tasks, and an order of magnitude larger than previous datasets. The
dataset contains 250K task-oriented grasps for 56 tasks and 191 objects along
with their RGB-D information. We take advantage of this new breadth and
diversity in the data and present the GCNGrasp framework which uses the
semantic knowledge of objects and tasks encoded in a knowledge graph to
generalize to new object instances, classes and even new tasks. Our framework
shows a significant improvement of around 12% on held-out settings compared to
baseline methods which do not use semantics. We demonstrate that our dataset
and model are applicable for the real world by executing task-oriented grasps
on a real robot on unknown objects. Code, data and supplementary video could be
found at https://sites.google.com/view/taskgrasp
Related papers
- Leverage Task Context for Object Affordance Ranking [57.59106517732223]
We build the first large-scale task-oriented affordance ranking dataset with 25 common tasks, over 50k images and more than 661k objects.
Results demonstrate the feasibility of the task context based affordance learning paradigm and the superiority of our model over state-of-the-art models in the fields of saliency ranking and multimodal object detection.
arXiv Detail & Related papers (2024-11-25T04:22:33Z) - Distribution Matching for Multi-Task Learning of Classification Tasks: a
Large-Scale Study on Faces & Beyond [62.406687088097605]
Multi-Task Learning (MTL) is a framework, where multiple related tasks are learned jointly and benefit from a shared representation space.
We show that MTL can be successful with classification tasks with little, or non-overlapping annotations.
We propose a novel approach, where knowledge exchange is enabled between the tasks via distribution matching.
arXiv Detail & Related papers (2024-01-02T14:18:11Z) - Few-Shot In-Context Imitation Learning via Implicit Graph Alignment [15.215659641228655]
We formulate imitation learning as a conditional alignment problem between graph representations of objects.
We show that this conditioning allows for in-context learning, where a robot can perform a task on a set of new objects immediately after the demonstrations.
arXiv Detail & Related papers (2023-10-18T18:26:01Z) - MetaGraspNet: A Large-Scale Benchmark Dataset for Vision-driven Robotic
Grasping via Physics-based Metaverse Synthesis [78.26022688167133]
We present a large-scale benchmark dataset for vision-driven robotic grasping via physics-based metaverse synthesis.
The proposed dataset contains 100,000 images and 25 different object types.
We also propose a new layout-weighted performance metric alongside the dataset for evaluating object detection and segmentation performance.
arXiv Detail & Related papers (2021-12-29T17:23:24Z) - Instance-Level Task Parameters: A Robust Multi-task Weighting Framework [17.639472693362926]
Recent works have shown that deep neural networks benefit from multi-task learning by learning a shared representation across several related tasks.
We let the training process dictate the optimal weighting of tasks for every instance in the dataset.
We conduct extensive experiments on SURREAL and CityScapes datasets, for human shape and pose estimation, depth estimation and semantic segmentation tasks.
arXiv Detail & Related papers (2021-06-11T02:35:42Z) - REGRAD: A Large-Scale Relational Grasp Dataset for Safe and
Object-Specific Robotic Grasping in Clutter [52.117388513480435]
We present a new dataset named regrad to sustain the modeling of relationships among objects and grasps.
Our dataset is collected in both forms of 2D images and 3D point clouds.
Users are free to import their own object models for the generation of as many data as they want.
arXiv Detail & Related papers (2021-04-29T05:31:21Z) - The IKEA ASM Dataset: Understanding People Assembling Furniture through
Actions, Objects and Pose [108.21037046507483]
IKEA ASM is a three million frame, multi-view, furniture assembly video dataset that includes depth, atomic actions, object segmentation, and human pose.
We benchmark prominent methods for video action recognition, object segmentation and human pose estimation tasks on this challenging dataset.
The dataset enables the development of holistic methods, which integrate multi-modal and multi-view data to better perform on these tasks.
arXiv Detail & Related papers (2020-07-01T11:34:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.