Skeletal Feature Compensation for Imitation Learning with Embodiment
Mismatch
- URL: http://arxiv.org/abs/2104.07810v1
- Date: Thu, 15 Apr 2021 22:50:48 GMT
- Title: Skeletal Feature Compensation for Imitation Learning with Embodiment
Mismatch
- Authors: Eddy Hudson, Garrett Warnell, Faraz Torabi, Peter Stone
- Abstract summary: SILEM is a proposed imitation learning technique that compensates for differences in skeletal features obtained from the learner and expert.
We create toy domains based on PyBullet's HalfCheetah and Ant to assess SILEM's benefits for this type of embodiment mismatch.
We also provide qualitative and quantitative results on more realistic problems -- teaching simulated humanoid agents to walk by observing human demonstrations.
- Score: 51.03498820458658
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Learning from demonstrations in the wild (e.g. YouTube videos) is a
tantalizing goal in imitation learning. However, for this goal to be achieved,
imitation learning algorithms must deal with the fact that the demonstrators
and learners may have bodies that differ from one another. This condition --
"embodiment mismatch" -- is ignored by many recent imitation learning
algorithms. Our proposed imitation learning technique, SILEM (\textbf{S}keletal
feature compensation for \textbf{I}mitation \textbf{L}earning with
\textbf{E}mbodiment \textbf{M}ismatch), addresses a particular type of
embodiment mismatch by introducing a learned affine transform to compensate for
differences in the skeletal features obtained from the learner and expert. We
create toy domains based on PyBullet's HalfCheetah and Ant to assess SILEM's
benefits for this type of embodiment mismatch. We also provide qualitative and
quantitative results on more realistic problems -- teaching simulated humanoid
agents, including Atlas from Boston Dynamics, to walk by observing human
demonstrations.
Related papers
- Boosting Semi-Supervised Scene Text Recognition via Viewing and Summarizing [71.29488677105127]
Existing scene text recognition (STR) methods struggle to recognize challenging texts, especially for artistic and severely distorted characters.
We propose a contrastive learning-based STR framework by leveraging synthetic and real unlabeled data without any human cost.
Our method achieves SOTA performance (94.7% and 70.9% average accuracy on common benchmarks and Union14M-Benchmark.
arXiv Detail & Related papers (2024-11-23T15:24:47Z) - Representation Alignment from Human Feedback for Cross-Embodiment Reward Learning from Mixed-Quality Demonstrations [8.71931996488953]
We study the problem of cross-embodiment inverse reinforcement learning, where we wish to learn a reward function from video demonstrations in one or more embodiments.
We analyze several techniques that leverage human feedback for representation learning and alignment to enable effective cross-embodiment learning.
arXiv Detail & Related papers (2024-08-10T18:24:14Z) - Contrastive Unlearning: A Contrastive Approach to Machine Unlearning [30.38966646250252]
We propose a contrastive unlearning framework, leveraging the concept of representation learning for more effective unlearning.
We show that contrastive unlearning achieves the best unlearning effects and efficiency with the lowest performance loss compared with the state-of-the-art algorithms.
arXiv Detail & Related papers (2024-01-19T02:16:30Z) - Sequential Causal Imitation Learning with Unobserved Confounders [82.22545916247269]
"Monkey see monkey do" is an age-old adage, referring to na"ive imitation without a deep understanding of a system's underlying mechanics.
This paper investigates the problem of causal imitation learning in sequential settings, where the imitator must make multiple decisions per episode.
arXiv Detail & Related papers (2022-08-12T13:53:23Z) - Partner-Assisted Learning for Few-Shot Image Classification [54.66864961784989]
Few-shot Learning has been studied to mimic human visual capabilities and learn effective models without the need of exhaustive human annotation.
In this paper, we focus on the design of training strategy to obtain an elemental representation such that the prototype of each novel class can be estimated from a few labeled samples.
We propose a two-stage training scheme, which first trains a partner encoder to model pair-wise similarities and extract features serving as soft-anchors, and then trains a main encoder by aligning its outputs with soft-anchors while attempting to maximize classification performance.
arXiv Detail & Related papers (2021-09-15T22:46:19Z) - Few-shot Visual Reasoning with Meta-analogical Contrastive Learning [141.2562447971]
We propose to solve a few-shot (or low-shot) visual reasoning problem, by resorting to analogical reasoning.
We extract structural relationships between elements in both domains, and enforce them to be as similar as possible with analogical learning.
We validate our method on RAVEN dataset, on which it outperforms state-of-the-art method, with larger gains when the training data is scarce.
arXiv Detail & Related papers (2020-07-23T14:00:34Z) - Adversarial Machine Learning in Network Intrusion Detection Systems [6.18778092044887]
We study the nature of the adversarial problem in Network Intrusion Detection Systems.
We use evolutionary computation (particle swarm optimization and genetic algorithm) and deep learning (generative adversarial networks) as tools for adversarial example generation.
Our work highlights the vulnerability of machine learning based NIDS in the face of adversarial perturbation.
arXiv Detail & Related papers (2020-04-23T19:47:43Z) - State-Only Imitation Learning for Dexterous Manipulation [63.03621861920732]
In this paper, we explore state-only imitation learning.
We train an inverse dynamics model and use it to predict actions for state-only demonstrations.
Our method performs on par with state-action approaches and considerably outperforms RL alone.
arXiv Detail & Related papers (2020-04-07T17:57:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.