Fast Online Adaptation in Robotics through Meta-Learning Embeddings of
Simulated Priors
- URL: http://arxiv.org/abs/2003.04663v2
- Date: Wed, 6 Jan 2021 19:59:40 GMT
- Title: Fast Online Adaptation in Robotics through Meta-Learning Embeddings of
Simulated Priors
- Authors: Rituraj Kaushik, Timoth\'ee Anne and Jean-Baptiste Mouret
- Abstract summary: In the real world, a robot might encounter any situation starting from motor failures to finding itself in a rocky terrain.
We show that FAMLE allows the robots to adapt to novel damages in significantly fewer time-steps than the baselines.
- Score: 3.4376560669160385
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Meta-learning algorithms can accelerate the model-based reinforcement
learning (MBRL) algorithms by finding an initial set of parameters for the
dynamical model such that the model can be trained to match the actual dynamics
of the system with only a few data-points. However, in the real world, a robot
might encounter any situation starting from motor failures to finding itself in
a rocky terrain where the dynamics of the robot can be significantly different
from one another. In this paper, first, we show that when meta-training
situations (the prior situations) have such diverse dynamics, using a single
set of meta-trained parameters as a starting point still requires a large
number of observations from the real system to learn a useful model of the
dynamics. Second, we propose an algorithm called FAMLE that mitigates this
limitation by meta-training several initial starting points (i.e., initial
parameters) for training the model and allows the robot to select the most
suitable starting point to adapt the model to the current situation with only a
few gradient steps. We compare FAMLE to MBRL, MBRL with a meta-trained model
with MAML, and model-free policy search algorithm PPO for various simulated and
real robotic tasks, and show that FAMLE allows the robots to adapt to novel
damages in significantly fewer time-steps than the baselines.
Related papers
- Learning Low-Dimensional Strain Models of Soft Robots by Looking at the Evolution of Their Shape with Application to Model-Based Control [2.058941610795796]
This paper introduces a streamlined method for learning low-dimensional, physics-based models.
We validate our approach through simulations with various planar soft manipulators.
Thanks to the capability of the method of generating physically compatible models, the learned models can be straightforwardly combined with model-based control policies.
arXiv Detail & Related papers (2024-10-31T18:37:22Z) - Data-Efficient Task Generalization via Probabilistic Model-based Meta
Reinforcement Learning [58.575939354953526]
PACOH-RL is a novel model-based Meta-Reinforcement Learning (Meta-RL) algorithm designed to efficiently adapt control policies to changing dynamics.
Existing Meta-RL methods require abundant meta-learning data, limiting their applicability in settings such as robotics.
Our experiment results demonstrate that PACOH-RL outperforms model-based RL and model-based Meta-RL baselines in adapting to new dynamic conditions.
arXiv Detail & Related papers (2023-11-13T18:51:57Z) - Physics-informed reinforcement learning via probabilistic co-adjustment
functions [3.6787556334630334]
We introduce co-kriging adjustments (CKA) and ridge regression adjustment (RRA) as novel ways to combine the advantages of both approaches.
Our adjustment methods are based on an auto-regressive AR1 co-kriging model that we integrate with GP priors.
arXiv Detail & Related papers (2023-09-11T12:10:19Z) - Domain Randomization for Robust, Affordable and Effective Closed-loop
Control of Soft Robots [10.977130974626668]
Soft robots are gaining popularity thanks to their intrinsic safety to contacts and adaptability.
We show how Domain Randomization (DR) can solve this problem by enhancing RL policies for soft robots.
We introduce a novel algorithmic extension to previous adaptive domain randomization methods for the automatic inference of dynamics parameters for deformable objects.
arXiv Detail & Related papers (2023-03-07T18:50:00Z) - Real-to-Sim: Predicting Residual Errors of Robotic Systems with Sparse
Data using a Learning-based Unscented Kalman Filter [65.93205328894608]
We learn the residual errors between a dynamic and/or simulator model and the real robot.
We show that with the learned residual errors, we can further close the reality gap between dynamic models, simulations, and actual hardware.
arXiv Detail & Related papers (2022-09-07T15:15:12Z) - Gradient-Based Trajectory Optimization With Learned Dynamics [80.41791191022139]
We use machine learning techniques to learn a differentiable dynamics model of the system from data.
We show that a neural network can model highly nonlinear behaviors accurately for large time horizons.
In our hardware experiments, we demonstrate that our learned model can represent complex dynamics for both the Spot and Radio-controlled (RC) car.
arXiv Detail & Related papers (2022-04-09T22:07:34Z) - Bayesian Meta-Learning for Few-Shot Policy Adaptation Across Robotic
Platforms [60.59764170868101]
Reinforcement learning methods can achieve significant performance but require a large amount of training data collected on the same robotic platform.
We formulate it as a few-shot meta-learning problem where the goal is to find a model that captures the common structure shared across different robotic platforms.
We experimentally evaluate our framework on a simulated reaching and a real-robot picking task using 400 simulated robots.
arXiv Detail & Related papers (2021-03-05T14:16:20Z) - MELD: Meta-Reinforcement Learning from Images via Latent State Models [109.1664295663325]
We develop an algorithm for meta-RL from images that performs inference in a latent state model to quickly acquire new skills.
MELD is the first meta-RL algorithm trained in a real-world robotic control setting from images.
arXiv Detail & Related papers (2020-10-26T23:50:30Z) - Information Theoretic Model Predictive Q-Learning [64.74041985237105]
We present a novel theoretical connection between information theoretic MPC and entropy regularized RL.
We develop a Q-learning algorithm that can leverage biased models.
arXiv Detail & Related papers (2019-12-31T00:29:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.