Adaptive Asynchronous Control Using Meta-learned Neural Ordinary
Differential Equations
- URL: http://arxiv.org/abs/2207.12062v5
- Date: Mon, 23 Oct 2023 10:02:22 GMT
- Title: Adaptive Asynchronous Control Using Meta-learned Neural Ordinary
Differential Equations
- Authors: Achkan Salehi, Steffen R\"uhl, Stephane Doncieux
- Abstract summary: Real-world robotics systems often present challenges that limit the applicability of model-based Reinforcement Learning and Control.
We propose a general framework that overcomes those difficulties by meta-learning adaptive dynamics models for continuous-time prediction and control.
We present evaluations in two different robot simulations and on a real industrial robot.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Model-based Reinforcement Learning and Control have demonstrated great
potential in various sequential decision making problem domains, including in
robotics settings. However, real-world robotics systems often present
challenges that limit the applicability of those methods. In particular, we
note two problems that jointly happen in many industrial systems: 1)
Irregular/asynchronous observations and actions and 2) Dramatic changes in
environment dynamics from an episode to another (e.g. varying payload inertial
properties). We propose a general framework that overcomes those difficulties
by meta-learning adaptive dynamics models for continuous-time prediction and
control. The proposed approach is task-agnostic and can be adapted to new tasks
in a straight-forward manner. We present evaluations in two different robot
simulations and on a real industrial robot.
Related papers
- Commonsense Reasoning for Legged Robot Adaptation with Vision-Language Models [81.55156507635286]
Legged robots are physically capable of navigating a diverse variety of environments and overcoming a wide range of obstructions.
Current learning methods often struggle with generalization to the long tail of unexpected situations without heavy human supervision.
We propose a system, VLM-Predictive Control (VLM-PC), combining two key components that we find to be crucial for eliciting on-the-fly, adaptive behavior selection.
arXiv Detail & Related papers (2024-07-02T21:00:30Z) - Adapt On-the-Go: Behavior Modulation for Single-Life Robot Deployment [92.48012013825988]
We study the problem of adapting on-the-fly to novel scenarios during deployment.
Our approach, RObust Autonomous Modulation (ROAM), introduces a mechanism based on the perceived value of pre-trained behaviors.
We demonstrate that ROAM enables a robot to adapt rapidly to changes in dynamics both in simulation and on a real Go1 quadruped.
arXiv Detail & Related papers (2023-11-02T08:22:28Z) - Domain Randomization for Robust, Affordable and Effective Closed-loop
Control of Soft Robots [10.977130974626668]
Soft robots are gaining popularity thanks to their intrinsic safety to contacts and adaptability.
We show how Domain Randomization (DR) can solve this problem by enhancing RL policies for soft robots.
We introduce a novel algorithmic extension to previous adaptive domain randomization methods for the automatic inference of dynamics parameters for deformable objects.
arXiv Detail & Related papers (2023-03-07T18:50:00Z) - Universal Morphology Control via Contextual Modulation [52.742056836818136]
Learning a universal policy across different robot morphologies can significantly improve learning efficiency and generalization in continuous control.
Existing methods utilize graph neural networks or transformers to handle heterogeneous state and action spaces across different morphologies.
We propose a hierarchical architecture to better model this dependency via contextual modulation.
arXiv Detail & Related papers (2023-02-22T00:04:12Z) - Avoiding Catastrophe: Active Dendrites Enable Multi-Task Learning in
Dynamic Environments [0.5277756703318046]
Key challenge for AI is to build embodied systems that operate in dynamically changing environments.
Standard deep learning systems often struggle in dynamic scenarios.
In this article we investigate biologically inspired architectures as solutions.
arXiv Detail & Related papers (2021-12-31T19:52:42Z) - Learning Reactive and Predictive Differentiable Controllers for
Switching Linear Dynamical Models [7.653542219337937]
We present a framework for learning composite dynamical behaviors from expert demonstrations.
We learn a switching linear dynamical model with contacts encoded in switching conditions as a close approximation of our system dynamics.
We then use discrete-time LQR as the differentiable policy class for data-efficient learning of control to develop a control strategy.
arXiv Detail & Related papers (2021-03-26T04:40:24Z) - Learning to Shift Attention for Motion Generation [55.61994201686024]
One challenge of motion generation using robot learning from demonstration techniques is that human demonstrations follow a distribution with multiple modes for one task query.
Previous approaches fail to capture all modes or tend to average modes of the demonstrations and thus generate invalid trajectories.
We propose a motion generation model with extrapolation ability to overcome this problem.
arXiv Detail & Related papers (2021-02-24T09:07:52Z) - Meta-Reinforcement Learning for Adaptive Motor Control in Changing Robot
Dynamics and Environments [3.5309638744466167]
This work developed a meta-learning approach that adapts the control policy on the fly to different changing conditions for robust locomotion.
The proposed method constantly updates the interaction model, samples feasible sequences of actions of estimated the state-action trajectories, and then applies the optimal actions to maximize the reward.
arXiv Detail & Related papers (2021-01-19T12:57:12Z) - Neural Dynamic Policies for End-to-End Sensorimotor Learning [51.24542903398335]
The current dominant paradigm in sensorimotor control, whether imitation or reinforcement learning, is to train policies directly in raw action spaces.
We propose Neural Dynamic Policies (NDPs) that make predictions in trajectory distribution space.
NDPs outperform the prior state-of-the-art in terms of either efficiency or performance across several robotic control tasks.
arXiv Detail & Related papers (2020-12-04T18:59:32Z) - ADAIL: Adaptive Adversarial Imitation Learning [11.270858993502705]
We present the ADaptive Adversarial Imitation Learning (ADAIL) algorithm for learning adaptive policies that can be transferred between environments of varying dynamics.
This is an important problem in robotic learning because in real world scenarios 1) reward functions are hard to obtain, 2) learned policies from one domain are difficult to deploy in another due to varying source to target domain statistics, and 3) collecting expert demonstrations in multiple environments where the dynamics are known and controlled is often infeasible.
arXiv Detail & Related papers (2020-08-23T06:11:00Z) - Learning to Control PDEs with Differentiable Physics [102.36050646250871]
We present a novel hierarchical predictor-corrector scheme which enables neural networks to learn to understand and control complex nonlinear physical systems over long time frames.
We demonstrate that our method successfully develops an understanding of complex physical systems and learns to control them for tasks involving PDEs.
arXiv Detail & Related papers (2020-01-21T11:58:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.