Learning Transferable Push Manipulation Skills in Novel Contexts
- URL: http://arxiv.org/abs/2007.14755v1
- Date: Wed, 29 Jul 2020 11:48:56 GMT
- Title: Learning Transferable Push Manipulation Skills in Novel Contexts
- Authors: Rhys Howard and Claudio Zito
- Abstract summary: We learn a parametric internal model for push interactions that enables a robot to predict the outcome of a physical interaction even in novel contexts.
We train on 2 objects for a total of 24,000 pushes in various conditions, and test on 6 objects for a total of 14,400 predicted push outcomes.
Our results show that both biased and unbiased predictors can reliably produce predictions in line with the outcomes of a carefully tuned physics simulator.
- Score: 3.1981440103815717
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper is concerned with learning transferable forward models for push
manipulation that can be applying to novel contexts and how to improve the
quality of prediction when critical information is available. We propose to
learn a parametric internal model for push interactions that, similar for
humans, enables a robot to predict the outcome of a physical interaction even
in novel contexts. Given a desired push action, humans are capable to identify
where to place their finger on a new object so to produce a predictable motion
of the object. We achieve the same behaviour by factorising the learning into
two parts. First, we learn a set of local contact models to represent the
geometrical relations between the robot pusher, the object, and the
environment. Then we learn a set of parametric local motion models to predict
how these contacts change throughout a push. The set of contact and motion
models represent our internal model. By adjusting the shapes of the
distributions over the physical parameters, we modify the internal model's
response. Uniform distributions yield to coarse estimates when no information
is available about the novel context (i.e. unbiased predictor). A more accurate
predictor can be learned for a specific environment/object pair (e.g. low
friction/high mass), i.e. biased predictor. The effectiveness of our approach
is shown in a simulated environment in which a Pioneer 3-DX robot needs to
predict a push outcome for a novel object, and we provide a proof of concept on
a real robot. We train on 2 objects (a cube and a cylinder) for a total of
24,000 pushes in various conditions, and test on 6 objects encompassing a
variety of shapes, sizes, and physical parameters for a total of 14,400
predicted push outcomes. Our results show that both biased and unbiased
predictors can reliably produce predictions in line with the outcomes of a
carefully tuned physics simulator.
Related papers
- RoboPack: Learning Tactile-Informed Dynamics Models for Dense Packing [38.97168020979433]
We introduce an approach that combines visual and tactile sensing for robotic manipulation by learning a neural, tactile-informed dynamics model.
Our proposed framework, RoboPack, employs a recurrent graph neural network to estimate object states.
We demonstrate our approach on a real robot equipped with a compliant Soft-Bubble tactile sensor on non-prehensile manipulation and dense packing tasks.
arXiv Detail & Related papers (2024-07-01T16:08:37Z) - Robot Learning with Sensorimotor Pre-training [98.7755895548928]
We present a self-supervised sensorimotor pre-training approach for robotics.
Our model, called RPT, is a Transformer that operates on sequences of sensorimotor tokens.
We find that sensorimotor pre-training consistently outperforms training from scratch, has favorable scaling properties, and enables transfer across different tasks, environments, and robots.
arXiv Detail & Related papers (2023-06-16T17:58:10Z) - Visual Affordance Prediction for Guiding Robot Exploration [56.17795036091848]
We develop an approach for learning visual affordances for guiding robot exploration.
We use a Transformer-based model to learn a conditional distribution in the latent embedding space of a VQ-VAE.
We show how the trained affordance model can be used for guiding exploration by acting as a goal-sampling distribution, during visual goal-conditioned policy learning in robotic manipulation.
arXiv Detail & Related papers (2023-05-28T17:53:09Z) - Combining Vision and Tactile Sensation for Video Prediction [0.0]
We investigate the impact of integrating tactile feedback into video prediction models for physical robot interactions.
We introduce two new datasets of robot pushing that use a magnetic-based tactile sensor for unsupervised learning.
Our results demonstrate that incorporating tactile feedback into video prediction models improves scene prediction accuracy and enhances the agent's perception of physical interactions.
arXiv Detail & Related papers (2023-04-21T18:02:15Z) - Continual Predictive Learning from Videos [100.27176974654559]
We study a new continual learning problem in the context of video prediction.
We propose the continual predictive learning (CPL) approach, which learns a mixture world model via predictive experience replay.
We construct two new benchmarks based on RoboNet and KTH, in which different tasks correspond to different physical robotic environments or human actions.
arXiv Detail & Related papers (2022-04-12T08:32:26Z) - Factored World Models for Zero-Shot Generalization in Robotic
Manipulation [7.258229016768018]
We learn to generalize over robotic pick-and-place tasks using object-factored world models.
We use a residual stack of graph neural networks that receive action information at multiple levels in both their node and edge neural networks.
We show that an ensemble of our models can be used to plan for tasks involving up to 12 pick and place actions using search.
arXiv Detail & Related papers (2022-02-10T21:26:11Z) - Dynamic Modeling of Hand-Object Interactions via Tactile Sensing [133.52375730875696]
In this work, we employ a high-resolution tactile glove to perform four different interactive activities on a diversified set of objects.
We build our model on a cross-modal learning framework and generate the labels using a visual processing pipeline to supervise the tactile model.
This work takes a step on dynamics modeling in hand-object interactions from dense tactile sensing.
arXiv Detail & Related papers (2021-09-09T16:04:14Z) - Few-Shot Visual Grounding for Natural Human-Robot Interaction [0.0]
We propose a software architecture that segments a target object from a crowded scene, indicated verbally by a human user.
At the core of our system, we employ a multi-modal deep neural network for visual grounding.
We evaluate the performance of the proposed model on real RGB-D data collected from public scene datasets.
arXiv Detail & Related papers (2021-03-17T15:24:02Z) - Object and Relation Centric Representations for Push Effect Prediction [18.990827725752496]
Pushing is an essential non-prehensile manipulation skill used for tasks ranging from pre-grasp manipulation to scene rearrangement.
We propose a graph neural network based framework for effect prediction and parameter estimation of pushing actions.
Our framework is validated both in real and simulated environments containing different shaped multi-part objects connected via different types of joints and objects with different masses.
arXiv Detail & Related papers (2021-02-03T15:09:12Z) - Predicting the Physical Dynamics of Unseen 3D Objects [65.49291702488436]
We focus on predicting the dynamics of 3D objects on a plane that have just been subjected to an impulsive force.
Our approach can generalize to object shapes and initial conditions that were unseen during training.
Our model can support training with data from both a physics engine or the real world.
arXiv Detail & Related papers (2020-01-16T06:27:59Z) - Learning Predictive Models From Observation and Interaction [137.77887825854768]
Learning predictive models from interaction with the world allows an agent, such as a robot, to learn about how the world works.
However, learning a model that captures the dynamics of complex skills represents a major challenge.
We propose a method to augment the training set with observational data of other agents, such as humans.
arXiv Detail & Related papers (2019-12-30T01:10:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.