A Road-map to Robot Task Execution with the Functional Object-Oriented
Network
- URL: http://arxiv.org/abs/2106.00158v1
- Date: Tue, 1 Jun 2021 00:43:04 GMT
- Title: A Road-map to Robot Task Execution with the Functional Object-Oriented
Network
- Authors: David Paulius, Alejandro Agostini, Yu Sun and Dongheui Lee
- Abstract summary: functional object-oriented network (FOON) is a knowledge graph representation for robots.
Taking the form of a bipartite graph, a FOON contains symbolic or high-level information that would be pertinent to a robot's understanding of its environment and tasks.
- Score: 77.93376696738409
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Following work on joint object-action representations, the functional
object-oriented network (FOON) was introduced as a knowledge graph
representation for robots. Taking the form of a bipartite graph, a FOON
contains symbolic or high-level information that would be pertinent to a
robot's understanding of its environment and tasks in a way that mirrors human
understanding of actions. In this work, we outline a road-map for future
development of FOON and its application in robotic systems for task planning as
well as knowledge acquisition from demonstration. We propose preliminary ideas
to show how a FOON can be created in a real-world scenario with a robot and
human teacher in a way that can jointly augment existing knowledge in a FOON
and teach a robot the skills it needs to replicate the demonstrated actions and
solve a given manipulation problem.
Related papers
- Task Tree Retrieval For Robotic Cooking [0.0]
This paper is based on developing different algorithms, which generate the task tree planning for the given goal node(recipe)
The knowledge representation of the dishes is called FOON. It contains the different objects and their between them with respective to the motion node.
arXiv Detail & Related papers (2023-11-27T20:41:21Z) - Human-oriented Representation Learning for Robotic Manipulation [64.59499047836637]
Humans inherently possess generalizable visual representations that empower them to efficiently explore and interact with the environments in manipulation tasks.
We formalize this idea through the lens of human-oriented multi-task fine-tuning on top of pre-trained visual encoders.
Our Task Fusion Decoder consistently improves the representation of three state-of-the-art visual encoders for downstream manipulation policy-learning.
arXiv Detail & Related papers (2023-10-04T17:59:38Z) - Learning Reward Functions for Robotic Manipulation by Observing Humans [92.30657414416527]
We use unlabeled videos of humans solving a wide range of manipulation tasks to learn a task-agnostic reward function for robotic manipulation policies.
The learned rewards are based on distances to a goal in an embedding space learned using a time-contrastive objective.
arXiv Detail & Related papers (2022-11-16T16:26:48Z) - Knowledge Retrieval using Functional Object-Oriented Network [0.0]
The functional object-oriented network (FOON) is a knowledge representation for symbolic task planning that takes the shape of a graph.
A graph retrieval methodology is shown to produce manipulation motion sequences from the FOON to accomplish a desired aim.
The outcomes are illustrated using motion sequences created by the FOON to complete the desired objectives in a simulated environment.
arXiv Detail & Related papers (2022-11-06T06:02:29Z) - Long-Horizon Planning and Execution with Functional Object-Oriented
Networks [79.94575713911189]
We introduce the idea of exploiting object-level knowledge as a FOON for task planning and execution.
Our approach automatically transforms FOON into PDDL and leverages off-the-shelf planners, action contexts, and robot skills.
We demonstrate our approach on long-horizon tasks in CoppeliaSim and show how learned action contexts can be extended to never-before-seen scenarios.
arXiv Detail & Related papers (2022-07-12T19:29:35Z) - Learning Generalizable Robotic Reward Functions from "In-The-Wild" Human
Videos [59.58105314783289]
Domain-agnostic Video Discriminator (DVD) learns multitask reward functions by training a discriminator to classify whether two videos are performing the same task.
DVD can generalize by virtue of learning from a small amount of robot data with a broad dataset of human videos.
DVD can be combined with visual model predictive control to solve robotic manipulation tasks on a real WidowX200 robot in an unseen environment from a single human demo.
arXiv Detail & Related papers (2021-03-31T05:25:05Z) - Axiom Learning and Belief Tracing for Transparent Decision Making in
Robotics [8.566457170664926]
A robot's ability to provide descriptions of its decisions and beliefs promotes effective collaboration with humans.
Our architecture couples the complementary strengths of non-monotonic logical reasoning, deep learning, and decision-tree induction.
During reasoning and learning, the architecture enables a robot to provide on-demand relational descriptions of its decisions, beliefs, and the outcomes of hypothetical actions.
arXiv Detail & Related papers (2020-10-20T22:09:17Z) - SAPIEN: A SimulAted Part-based Interactive ENvironment [77.4739790629284]
SAPIEN is a realistic and physics-rich simulated environment that hosts a large-scale set for articulated objects.
We evaluate state-of-the-art vision algorithms for part detection and motion attribute recognition as well as demonstrate robotic interaction tasks.
arXiv Detail & Related papers (2020-03-19T00:11:34Z) - Understanding Contexts Inside Robot and Human Manipulation Tasks through
a Vision-Language Model and Ontology System in a Video Stream [4.450615100675747]
We present a vision dataset under a strictly constrained knowledge domain for both robot and human manipulations.
We propose a scheme to generate a combination of visual attentions and an evolving knowledge graph filled with commonsense knowledge.
The proposed scheme allows the robot to mimic human-like intentional behaviors by watching real-time videos.
arXiv Detail & Related papers (2020-03-02T19:48:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.