Learning Perceptual Concepts by Bootstrapping from Human Queries
- URL: http://arxiv.org/abs/2111.05251v1
- Date: Tue, 9 Nov 2021 16:43:46 GMT
- Title: Learning Perceptual Concepts by Bootstrapping from Human Queries
- Authors: Andreea Bobu, Chris Paxton, Wei Yang, Balakumar Sundaralingam, Yu-Wei
Chao, Maya Cakmak, Dieter Fox
- Abstract summary: We propose a new approach whereby the robot learns a low-dimensional variant of the concept and uses it to generate a larger data set for learning the concept in the high-dimensional space.
This lets it take advantage of semantically meaningful privileged information only accessible at training time, like object poses and bounding boxes, that allows for richer human interaction to speed up learning.
- Score: 41.07749131023931
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Robots need to be able to learn concepts from their users in order to adapt
their capabilities to each user's unique task. But when the robot operates on
high-dimensional inputs, like images or point clouds, this is impractical: the
robot needs an unrealistic amount of human effort to learn the new concept. To
address this challenge, we propose a new approach whereby the robot learns a
low-dimensional variant of the concept and uses it to generate a larger data
set for learning the concept in the high-dimensional space. This lets it take
advantage of semantically meaningful privileged information only accessible at
training time, like object poses and bounding boxes, that allows for richer
human interaction to speed up learning. We evaluate our approach by learning
prepositional concepts that describe object state or multi-object
relationships, like above, near, or aligned, which are key to user
specification of task goals and execution constraints for robots. Using a
simulated human, we show that our approach improves sample complexity when
compared to learning concepts directly in the high-dimensional space. We also
demonstrate the utility of the learned concepts in motion planning tasks on a
7-DoF Franka Panda robot.
Related papers
- Human-oriented Representation Learning for Robotic Manipulation [64.59499047836637]
Humans inherently possess generalizable visual representations that empower them to efficiently explore and interact with the environments in manipulation tasks.
We formalize this idea through the lens of human-oriented multi-task fine-tuning on top of pre-trained visual encoders.
Our Task Fusion Decoder consistently improves the representation of three state-of-the-art visual encoders for downstream manipulation policy-learning.
arXiv Detail & Related papers (2023-10-04T17:59:38Z) - Affordances from Human Videos as a Versatile Representation for Robotics [31.248842798600606]
We train a visual affordance model that estimates where and how in the scene a human is likely to interact.
The structure of these behavioral affordances directly enables the robot to perform many complex tasks.
We show the efficacy of our approach, which we call VRB, across 4 real world environments, over 10 different tasks, and 2 robotic platforms operating in the wild.
arXiv Detail & Related papers (2023-04-17T17:59:34Z) - Self-Improving Robots: End-to-End Autonomous Visuomotor Reinforcement
Learning [54.636562516974884]
In imitation and reinforcement learning, the cost of human supervision limits the amount of data that robots can be trained on.
In this work, we propose MEDAL++, a novel design for self-improving robotic systems.
The robot autonomously practices the task by learning to both do and undo the task, simultaneously inferring the reward function from the demonstrations.
arXiv Detail & Related papers (2023-03-02T18:51:38Z) - Learning Reward Functions for Robotic Manipulation by Observing Humans [92.30657414416527]
We use unlabeled videos of humans solving a wide range of manipulation tasks to learn a task-agnostic reward function for robotic manipulation policies.
The learned rewards are based on distances to a goal in an embedding space learned using a time-contrastive objective.
arXiv Detail & Related papers (2022-11-16T16:26:48Z) - Actionable Models: Unsupervised Offline Reinforcement Learning of
Robotic Skills [93.12417203541948]
We propose the objective of learning a functional understanding of the environment by learning to reach any goal state in a given dataset.
We find that our method can operate on high-dimensional camera images and learn a variety of skills on real robots that generalize to previously unseen scenes and objects.
arXiv Detail & Related papers (2021-04-15T20:10:11Z) - Learning Generalizable Robotic Reward Functions from "In-The-Wild" Human
Videos [59.58105314783289]
Domain-agnostic Video Discriminator (DVD) learns multitask reward functions by training a discriminator to classify whether two videos are performing the same task.
DVD can generalize by virtue of learning from a small amount of robot data with a broad dataset of human videos.
DVD can be combined with visual model predictive control to solve robotic manipulation tasks on a real WidowX200 robot in an unseen environment from a single human demo.
arXiv Detail & Related papers (2021-03-31T05:25:05Z) - The State of Lifelong Learning in Service Robots: Current Bottlenecks in
Object Perception and Manipulation [3.7858180627124463]
State-of-the-art continues to improve to make a proper coupling between object perception and manipulation.
In most of the cases, robots are able to recognize various objects, and quickly plan a collision-free trajectory to grasp a target object.
In such environments, no matter how extensive the training data used for batch learning, a robot will always face new objects.
apart from robot self-learning, non-expert users could interactively guide the process of experience acquisition.
arXiv Detail & Related papers (2020-03-18T11:00:55Z) - A Model of Fast Concept Inference with Object-Factorized Cognitive
Programs [3.4763296976688443]
We present an algorithm that emulates the human cognitives of object factorization and sub-goaling, allowing human-level inference speed, improving accuracy, and making the output more explainable.
arXiv Detail & Related papers (2020-02-10T18:48:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.