Metric-Based Imitation Learning Between Two Dissimilar Anthropomorphic
Robotic Arms
- URL: http://arxiv.org/abs/2003.02638v1
- Date: Tue, 25 Feb 2020 19:47:19 GMT
- Title: Metric-Based Imitation Learning Between Two Dissimilar Anthropomorphic
Robotic Arms
- Authors: Marcus Ebner von Eschenbach, Binyamin Manela, Jan Peters, Armin Biess
- Abstract summary: One major challenge in imitation learning is the correspondence problem.
We introduce a distance measure between dissimilar embodiments.
We find that the measure is well suited for describing the similarity between embodiments and for learning imitation policies by distance.
- Score: 29.08134072341867
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The development of autonomous robotic systems that can learn from human
demonstrations to imitate a desired behavior - rather than being manually
programmed - has huge technological potential. One major challenge in imitation
learning is the correspondence problem: how to establish corresponding states
and actions between expert and learner, when the embodiments of the agents are
different (morphology, dynamics, degrees of freedom, etc.). Many existing
approaches in imitation learning circumvent the correspondence problem, for
example, kinesthetic teaching or teleoperation, which are performed on the
robot. In this work we explicitly address the correspondence problem by
introducing a distance measure between dissimilar embodiments. This measure is
then used as a loss function for static pose imitation and as a feedback signal
within a model-free deep reinforcement learning framework for dynamic movement
imitation between two anthropomorphic robotic arms in simulation. We find that
the measure is well suited for describing the similarity between embodiments
and for learning imitation policies by distance minimization.
Related papers
- DiffGen: Robot Demonstration Generation via Differentiable Physics Simulation, Differentiable Rendering, and Vision-Language Model [72.66465487508556]
DiffGen is a novel framework that integrates differentiable physics simulation, differentiable rendering, and a vision-language model.
It can generate realistic robot demonstrations by minimizing the distance between the embedding of the language instruction and the embedding of the simulated observation.
Experiments demonstrate that with DiffGen, we could efficiently and effectively generate robot data with minimal human effort or training time.
arXiv Detail & Related papers (2024-05-12T15:38:17Z) - Real-time Addressee Estimation: Deployment of a Deep-Learning Model on
the iCub Robot [52.277579221741746]
Addressee Estimation is a skill essential for social robots to interact smoothly with humans.
Inspired by human perceptual skills, a deep-learning model for Addressee Estimation is designed, trained, and deployed on an iCub robot.
The study presents the procedure of such implementation and the performance of the model deployed in real-time human-robot interaction.
arXiv Detail & Related papers (2023-11-09T13:01:21Z) - DiAReL: Reinforcement Learning with Disturbance Awareness for Robust
Sim2Real Policy Transfer in Robot Control [0.0]
Delayed Markov decision processes fulfill the Markov property by augmenting the state space of agents with a finite time window of recently committed actions.
We introduce a disturbance-augmented Markov decision process in delayed settings as a novel representation to incorporate disturbance estimation in training on-policy reinforcement learning algorithms.
arXiv Detail & Related papers (2023-06-15T10:11:38Z) - Universal Morphology Control via Contextual Modulation [52.742056836818136]
Learning a universal policy across different robot morphologies can significantly improve learning efficiency and generalization in continuous control.
Existing methods utilize graph neural networks or transformers to handle heterogeneous state and action spaces across different morphologies.
We propose a hierarchical architecture to better model this dependency via contextual modulation.
arXiv Detail & Related papers (2023-02-22T00:04:12Z) - Navigating to Objects in the Real World [76.1517654037993]
We present a large-scale empirical study of semantic visual navigation methods comparing methods from classical, modular, and end-to-end learning approaches.
We find that modular learning works well in the real world, attaining a 90% success rate.
In contrast, end-to-end learning does not, dropping from 77% simulation to 23% real-world success rate due to a large image domain gap between simulation and reality.
arXiv Detail & Related papers (2022-12-02T01:10:47Z) - Coarse-to-Fine Imitation Learning: Robot Manipulation from a Single
Demonstration [8.57914821832517]
We introduce a simple new method for visual imitation learning, which allows a novel robot manipulation task to be learned from a single human demonstration.
Our method models imitation learning as a state estimation problem, with the state defined as the end-effector's pose.
At test time, the end-effector moves to the estimated state through a linear path, at which point the original demonstration's end-effector velocities are simply replayed.
arXiv Detail & Related papers (2021-05-13T16:36:55Z) - Learning to Shift Attention for Motion Generation [55.61994201686024]
One challenge of motion generation using robot learning from demonstration techniques is that human demonstrations follow a distribution with multiple modes for one task query.
Previous approaches fail to capture all modes or tend to average modes of the demonstrations and thus generate invalid trajectories.
We propose a motion generation model with extrapolation ability to overcome this problem.
arXiv Detail & Related papers (2021-02-24T09:07:52Z) - Learning Cross-Domain Correspondence for Control with Dynamics
Cycle-Consistency [60.39133304370604]
We learn to align dynamic robot behavior across two domains using a cycle-consistency constraint.
Our framework is able to align uncalibrated monocular video of a real robot arm to dynamic state-action trajectories of a simulated arm without paired data.
arXiv Detail & Related papers (2020-12-17T18:22:25Z) - Robotic self-representation improves manipulation skills and transfer
learning [14.863872352905629]
We develop a model that learns bidirectional action-effect associations to encode the representations of body schema and the peripersonal space from multisensory information.
We demonstrate that this approach significantly stabilizes the learning-based problem-solving under noisy conditions and that it improves transfer learning of robotic manipulation skills.
arXiv Detail & Related papers (2020-11-13T16:04:58Z) - Language-Conditioned Imitation Learning for Robot Manipulation Tasks [39.40937105264774]
We introduce a method for incorporating unstructured natural language into imitation learning.
At training time, the expert can provide demonstrations along with verbal descriptions in order to describe the underlying intent.
The training process then interrelates these two modalities to encode the correlations between language, perception, and motion.
The resulting language-conditioned visuomotor policies can be conditioned at runtime on new human commands and instructions.
arXiv Detail & Related papers (2020-10-22T21:49:08Z) - ContactNets: Learning Discontinuous Contact Dynamics with Smooth,
Implicit Representations [4.8986598953553555]
Our method learns parameterizations of inter-body signed distance and contact-frame Jacobians.
Our method can predict realistic impact, non-penetration, and stiction when trained on 60 seconds of real-world data.
arXiv Detail & Related papers (2020-09-23T14:51:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.