Model-Based Meta-Reinforcement Learning for Flight with Suspended
Payloads
- URL: http://arxiv.org/abs/2004.11345v2
- Date: Tue, 2 Feb 2021 06:32:03 GMT
- Title: Model-Based Meta-Reinforcement Learning for Flight with Suspended
Payloads
- Authors: Suneel Belkhale, Rachel Li, Gregory Kahn, Rowan McAllister, Roberto
Calandra, Sergey Levine
- Abstract summary: Transporting suspended payloads is challenging for autonomous aerial vehicles.
We propose a meta-learning approach that "learns how to learn" models of altered dynamics within seconds of post-connection flight data.
- Score: 69.21503033239985
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Transporting suspended payloads is challenging for autonomous aerial vehicles
because the payload can cause significant and unpredictable changes to the
robot's dynamics. These changes can lead to suboptimal flight performance or
even catastrophic failure. Although adaptive control and learning-based methods
can in principle adapt to changes in these hybrid robot-payload systems, rapid
mid-flight adaptation to payloads that have a priori unknown physical
properties remains an open problem. We propose a meta-learning approach that
"learns how to learn" models of altered dynamics within seconds of
post-connection flight data. Our experiments demonstrate that our online
adaptation approach outperforms non-adaptive methods on a series of challenging
suspended payload transportation tasks. Videos and other supplemental material
are available on our website: https://sites.google.com/view/meta-rl-for-flight
Related papers
- First, Learn What You Don't Know: Active Information Gathering for Driving at the Limits of Handling [38.468291768795865]
In unstable systems, online adaptation may not be fast enough to ensure reliable simultaneous learning and control.
We present a Bayesian meta-learning MPC framework to enable rapid online adaptation.
Experiments on a Toyota Supra show that the framework enables reliable control in dynamic drifting maneuvers.
arXiv Detail & Related papers (2024-10-31T18:02:30Z) - Physics Enhanced Residual Policy Learning (PERPL) for safety cruising in mixed traffic platooning under actuator and communication delay [8.172286651098027]
Linear control models have gained extensive application in vehicle control due to their simplicity, ease of use, and support for stability analysis.
Reinforcement learning (RL) models, on the other hand, offer adaptability but suffer from a lack of interpretability and generalization capabilities.
This paper aims to develop a family of RL-based controllers enhanced by physics-informed policies.
arXiv Detail & Related papers (2024-09-23T23:02:34Z) - Model-Based Reinforcement Learning with Multi-Task Offline Pretraining [59.82457030180094]
We present a model-based RL method that learns to transfer potentially useful dynamics and action demonstrations from offline data to a novel task.
The main idea is to use the world models not only as simulators for behavior learning but also as tools to measure the task relevance.
We demonstrate the advantages of our approach compared with the state-of-the-art methods in Meta-World and DeepMind Control Suite.
arXiv Detail & Related papers (2023-06-06T02:24:41Z) - OSCAR: Data-Driven Operational Space Control for Adaptive and Robust
Robot Manipulation [50.59541802645156]
Operational Space Control (OSC) has been used as an effective task-space controller for manipulation.
We propose OSC for Adaptation and Robustness (OSCAR), a data-driven variant of OSC that compensates for modeling errors.
We evaluate our method on a variety of simulated manipulation problems, and find substantial improvements over an array of controller baselines.
arXiv Detail & Related papers (2021-10-02T01:21:38Z) - Bayesian Meta-Learning for Few-Shot Policy Adaptation Across Robotic
Platforms [60.59764170868101]
Reinforcement learning methods can achieve significant performance but require a large amount of training data collected on the same robotic platform.
We formulate it as a few-shot meta-learning problem where the goal is to find a model that captures the common structure shared across different robotic platforms.
We experimentally evaluate our framework on a simulated reaching and a real-robot picking task using 400 simulated robots.
arXiv Detail & Related papers (2021-03-05T14:16:20Z) - Meta-Learning-Based Robust Adaptive Flight Control Under Uncertain Wind
Conditions [13.00214468719929]
Realtime model learning is challenging for complex dynamical systems, such as drones flying in variable wind conditions.
We propose an online composite adaptation method that treats outputs from a deep neural network as a set of basis functions.
We validate our approach by flying a drone in an open air wind tunnel under varying wind conditions and along challenging trajectories.
arXiv Detail & Related papers (2021-03-02T18:43:59Z) - Meta-Reinforcement Learning for Adaptive Motor Control in Changing Robot
Dynamics and Environments [3.5309638744466167]
This work developed a meta-learning approach that adapts the control policy on the fly to different changing conditions for robust locomotion.
The proposed method constantly updates the interaction model, samples feasible sequences of actions of estimated the state-action trajectories, and then applies the optimal actions to maximize the reward.
arXiv Detail & Related papers (2021-01-19T12:57:12Z) - Never Stop Learning: The Effectiveness of Fine-Tuning in Robotic
Reinforcement Learning [109.77163932886413]
We show how to adapt vision-based robotic manipulation policies to new variations by fine-tuning via off-policy reinforcement learning.
This adaptation uses less than 0.2% of the data necessary to learn the task from scratch.
We find that our approach of adapting pre-trained policies leads to substantial performance gains over the course of fine-tuning.
arXiv Detail & Related papers (2020-04-21T17:57:04Z) - Rapidly Adaptable Legged Robots via Evolutionary Meta-Learning [65.88200578485316]
We present a new meta-learning method that allows robots to quickly adapt to changes in dynamics.
Our method significantly improves adaptation to changes in dynamics in high noise settings.
We validate our approach on a quadruped robot that learns to walk while subject to changes in dynamics.
arXiv Detail & Related papers (2020-03-02T22:56:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.