Few-Shot Learning of Force-Based Motions From Demonstration Through
Pre-training of Haptic Representation
- URL: http://arxiv.org/abs/2309.04640v1
- Date: Fri, 8 Sep 2023 23:42:59 GMT
- Title: Few-Shot Learning of Force-Based Motions From Demonstration Through
Pre-training of Haptic Representation
- Authors: Marina Y. Aoyama, Jo\~ao Moura, Namiko Saito, Sethu Vijayakumar
- Abstract summary: Existing Learning from Demonstration (LfD) approaches require a large number of costly human demonstrations.
Our proposed semi-supervised LfD approach decouples the learnt model into an haptic representation encoder and a motion generation decoder.
This enables us to pre-train the first using large amount of unsupervised data, easily accessible, while using few-shot LfD to train the second.
We validate the motion generated by our semi-supervised LfD model on the physical robot hardware using the KUKA iiwa robot arm.
- Score: 10.553635668779911
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In many contact-rich tasks, force sensing plays an essential role in adapting
the motion to the physical properties of the manipulated object. To enable
robots to capture the underlying distribution of object properties necessary
for generalising learnt manipulation tasks to unseen objects, existing Learning
from Demonstration (LfD) approaches require a large number of costly human
demonstrations. Our proposed semi-supervised LfD approach decouples the learnt
model into an haptic representation encoder and a motion generation decoder.
This enables us to pre-train the first using large amount of unsupervised data,
easily accessible, while using few-shot LfD to train the second, leveraging the
benefits of learning skills from humans. We validate the approach on the wiping
task using sponges with different stiffness and surface friction. Our results
demonstrate that pre-training significantly improves the ability of the LfD
model to recognise physical properties and generate desired wiping motions for
unseen sponges, outperforming the LfD method without pre-training. We validate
the motion generated by our semi-supervised LfD model on the physical robot
hardware using the KUKA iiwa robot arm. We also validate that the haptic
representation encoder, pre-trained in simulation, captures the properties of
real objects, explaining its contribution to improving the generalisation of
the downstream task.
Related papers
- Affordance-Guided Reinforcement Learning via Visual Prompting [51.361977466993345]
We study rewards shaped by vision-language models (VLMs) to define dense rewards for robotic learning.
On a real-world manipulation task specified by natural language description, we find that these rewards improve the sample efficiency of autonomous RL.
arXiv Detail & Related papers (2024-07-14T21:41:29Z) - Conditional Neural Expert Processes for Learning Movement Primitives from Demonstration [1.9336815376402723]
Conditional Neural Expert Processes (CNEP) learns to assign demonstrations from different modes to distinct expert networks.
CNEP does not require supervision on which mode the trajectories belong to.
Our system is capable of on-the-fly adaptation to environmental changes via an online conditioning mechanism.
arXiv Detail & Related papers (2024-02-13T12:52:02Z) - Modular Neural Network Policies for Learning In-Flight Object Catching
with a Robot Hand-Arm System [55.94648383147838]
We present a modular framework designed to enable a robot hand-arm system to learn how to catch flying objects.
Our framework consists of five core modules: (i) an object state estimator that learns object trajectory prediction, (ii) a catching pose quality network that learns to score and rank object poses for catching, (iii) a reaching control policy trained to move the robot hand to pre-catch poses, and (iv) a grasping control policy trained to perform soft catching motions.
We conduct extensive evaluations of our framework in simulation for each module and the integrated system, to demonstrate high success rates of in-flight
arXiv Detail & Related papers (2023-12-21T16:20:12Z) - Movement Primitive Diffusion: Learning Gentle Robotic Manipulation of Deformable Objects [14.446751610174868]
Movement Primitive Diffusion (MPD) is a novel method for imitation learning (IL) in robot-assisted surgery.
MPD combines the versatility of diffusion-based imitation learning (DIL) with the high-quality motion generation capabilities of Probabilistic Dynamic Movement Primitives (ProDMPs)
We evaluate MPD across various simulated and real world robotic tasks on both state and image observations.
arXiv Detail & Related papers (2023-12-15T18:24:28Z) - Robotic Handling of Compliant Food Objects by Robust Learning from
Demonstration [79.76009817889397]
We propose a robust learning policy based on Learning from Demonstration (LfD) for robotic grasping of food compliant objects.
We present an LfD learning policy that automatically removes inconsistent demonstrations, and estimates the teacher's intended policy.
The proposed approach has a vast range of potential applications in the aforementioned industry sectors.
arXiv Detail & Related papers (2023-09-22T13:30:26Z) - SoftGPT: Learn Goal-oriented Soft Object Manipulation Skills by
Generative Pre-trained Heterogeneous Graph Transformer [34.86946655775187]
Soft object manipulation tasks in domestic scenes pose a significant challenge for existing robotic skill learning techniques.
We propose a pre-trained soft object manipulation skill learning model, namely SoftGPT, that is trained using large amounts of exploration data.
For each downstream task, a goal-oriented policy agent is trained to predict the subsequent actions, and SoftGPT generates the consequences.
arXiv Detail & Related papers (2023-06-22T05:48:22Z) - ASE: Large-Scale Reusable Adversarial Skill Embeddings for Physically
Simulated Characters [123.88692739360457]
General-purpose motor skills enable humans to perform complex tasks.
These skills also provide powerful priors for guiding their behaviors when learning new tasks.
We present a framework for learning versatile and reusable skill embeddings for physically simulated characters.
arXiv Detail & Related papers (2022-05-04T06:13:28Z) - Reinforcement Learning with Action-Free Pre-Training from Videos [95.25074614579646]
We introduce a framework that learns representations useful for understanding the dynamics via generative pre-training on videos.
Our framework significantly improves both final performances and sample-efficiency of vision-based reinforcement learning.
arXiv Detail & Related papers (2022-03-25T19:44:09Z) - Hierarchical Few-Shot Imitation with Skill Transition Models [66.81252581083199]
Few-shot Imitation with Skill Transition Models (FIST) is an algorithm that extracts skills from offline data and utilizes them to generalize to unseen tasks.
We show that FIST is capable of generalizing to new tasks and substantially outperforms prior baselines in navigation experiments.
arXiv Detail & Related papers (2021-07-19T15:56:01Z) - Understanding Multi-Modal Perception Using Behavioral Cloning for
Peg-In-a-Hole Insertion Tasks [21.275342989110978]
In this paper, we investigate the merits of multiple sensor modalities when combined to learn a controller for real world assembly operation tasks.
We propose a multi-step-ahead loss function to improve the performance of the behavioral cloning method.
arXiv Detail & Related papers (2020-07-22T19:46:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.