Few-Shot Learning of Force-Based Motions From Demonstration Through
Pre-training of Haptic Representation
- URL: http://arxiv.org/abs/2309.04640v1
- Date: Fri, 8 Sep 2023 23:42:59 GMT
- Title: Few-Shot Learning of Force-Based Motions From Demonstration Through
Pre-training of Haptic Representation
- Authors: Marina Y. Aoyama, Jo\~ao Moura, Namiko Saito, Sethu Vijayakumar
- Abstract summary: Existing Learning from Demonstration (LfD) approaches require a large number of costly human demonstrations.
Our proposed semi-supervised LfD approach decouples the learnt model into an haptic representation encoder and a motion generation decoder.
This enables us to pre-train the first using large amount of unsupervised data, easily accessible, while using few-shot LfD to train the second.
We validate the motion generated by our semi-supervised LfD model on the physical robot hardware using the KUKA iiwa robot arm.
- Score: 10.553635668779911
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In many contact-rich tasks, force sensing plays an essential role in adapting
the motion to the physical properties of the manipulated object. To enable
robots to capture the underlying distribution of object properties necessary
for generalising learnt manipulation tasks to unseen objects, existing Learning
from Demonstration (LfD) approaches require a large number of costly human
demonstrations. Our proposed semi-supervised LfD approach decouples the learnt
model into an haptic representation encoder and a motion generation decoder.
This enables us to pre-train the first using large amount of unsupervised data,
easily accessible, while using few-shot LfD to train the second, leveraging the
benefits of learning skills from humans. We validate the approach on the wiping
task using sponges with different stiffness and surface friction. Our results
demonstrate that pre-training significantly improves the ability of the LfD
model to recognise physical properties and generate desired wiping motions for
unseen sponges, outperforming the LfD method without pre-training. We validate
the motion generated by our semi-supervised LfD model on the physical robot
hardware using the KUKA iiwa robot arm. We also validate that the haptic
representation encoder, pre-trained in simulation, captures the properties of
real objects, explaining its contribution to improving the generalisation of
the downstream task.
Related papers
- Latent Action Pretraining from Videos [156.88613023078778]
We introduce Latent Action Pretraining for general Action models (LAPA)
LAPA is an unsupervised method for pretraining Vision-Language-Action (VLA) models without ground-truth robot action labels.
We propose a method to learn from internet-scale videos that do not have robot action labels.
arXiv Detail & Related papers (2024-10-15T16:28:09Z) - ReinDiffuse: Crafting Physically Plausible Motions with Reinforced Diffusion Model [9.525806425270428]
We present emphReinDiffuse that combines reinforcement learning with motion diffusion model to generate physically credible human motions.
Our method adapts Motion Diffusion Model to output a parameterized distribution of actions, making them compatible with reinforcement learning paradigms.
Our approach outperforms existing state-of-the-art models on two major datasets, HumanML3D and KIT-ML.
arXiv Detail & Related papers (2024-10-09T16:24:11Z) - Learning Gentle Grasping from Human-Free Force Control Demonstration [4.08734863805696]
We propose an approach for learning grasping from ideal force control demonstrations to achieve similar performance of human hands with limited data size.
Our approach utilizes objects with known contact characteristics to automatically generate reference force curves without human demonstrations.
The described method can be effectively applied in vision-based tactile sensors and enables gentle and stable grasping of objects from the ground.
arXiv Detail & Related papers (2024-09-16T15:14:53Z) - Affordance-Guided Reinforcement Learning via Visual Prompting [51.361977466993345]
Keypoint-based Affordance Guidance for Improvements (KAGI) is a method leveraging rewards shaped by vision-language models (VLMs) for autonomous RL.
On real-world manipulation tasks specified by natural language descriptions, KAGI improves the sample efficiency of autonomous RL and enables successful task completion in 20K online fine-tuning steps.
arXiv Detail & Related papers (2024-07-14T21:41:29Z) - Conditional Neural Expert Processes for Learning Movement Primitives from Demonstration [1.9336815376402723]
Conditional Neural Expert Processes (CNEP) learns to assign demonstrations from different modes to distinct expert networks.
CNEP does not require supervision on which mode the trajectories belong to.
Our system is capable of on-the-fly adaptation to environmental changes via an online conditioning mechanism.
arXiv Detail & Related papers (2024-02-13T12:52:02Z) - Movement Primitive Diffusion: Learning Gentle Robotic Manipulation of Deformable Objects [14.446751610174868]
Movement Primitive Diffusion (MPD) is a novel method for imitation learning (IL) in robot-assisted surgery.
MPD combines the versatility of diffusion-based imitation learning (DIL) with the high-quality motion generation capabilities of Probabilistic Dynamic Movement Primitives (ProDMPs)
We evaluate MPD across various simulated and real world robotic tasks on both state and image observations.
arXiv Detail & Related papers (2023-12-15T18:24:28Z) - Robotic Handling of Compliant Food Objects by Robust Learning from
Demonstration [79.76009817889397]
We propose a robust learning policy based on Learning from Demonstration (LfD) for robotic grasping of food compliant objects.
We present an LfD learning policy that automatically removes inconsistent demonstrations, and estimates the teacher's intended policy.
The proposed approach has a vast range of potential applications in the aforementioned industry sectors.
arXiv Detail & Related papers (2023-09-22T13:30:26Z) - SoftGPT: Learn Goal-oriented Soft Object Manipulation Skills by
Generative Pre-trained Heterogeneous Graph Transformer [34.86946655775187]
Soft object manipulation tasks in domestic scenes pose a significant challenge for existing robotic skill learning techniques.
We propose a pre-trained soft object manipulation skill learning model, namely SoftGPT, that is trained using large amounts of exploration data.
For each downstream task, a goal-oriented policy agent is trained to predict the subsequent actions, and SoftGPT generates the consequences.
arXiv Detail & Related papers (2023-06-22T05:48:22Z) - ASE: Large-Scale Reusable Adversarial Skill Embeddings for Physically
Simulated Characters [123.88692739360457]
General-purpose motor skills enable humans to perform complex tasks.
These skills also provide powerful priors for guiding their behaviors when learning new tasks.
We present a framework for learning versatile and reusable skill embeddings for physically simulated characters.
arXiv Detail & Related papers (2022-05-04T06:13:28Z) - Reinforcement Learning with Action-Free Pre-Training from Videos [95.25074614579646]
We introduce a framework that learns representations useful for understanding the dynamics via generative pre-training on videos.
Our framework significantly improves both final performances and sample-efficiency of vision-based reinforcement learning.
arXiv Detail & Related papers (2022-03-25T19:44:09Z) - Hierarchical Few-Shot Imitation with Skill Transition Models [66.81252581083199]
Few-shot Imitation with Skill Transition Models (FIST) is an algorithm that extracts skills from offline data and utilizes them to generalize to unseen tasks.
We show that FIST is capable of generalizing to new tasks and substantially outperforms prior baselines in navigation experiments.
arXiv Detail & Related papers (2021-07-19T15:56:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.