Robust, High-Rate Trajectory Tracking on Insect-Scale Soft-Actuated
Aerial Robots with Deep-Learned Tube MPC
- URL: http://arxiv.org/abs/2209.10007v1
- Date: Tue, 20 Sep 2022 21:30:16 GMT
- Title: Robust, High-Rate Trajectory Tracking on Insect-Scale Soft-Actuated
Aerial Robots with Deep-Learned Tube MPC
- Authors: Andrea Tagliabue (1), Yi-Hsuan Hsiao (2), Urban Fasel (3), J. Nathan
Kutz (4), Steven L. Brunton (5), YuFeng Chen (2) and Jonathan P. How (1) ((1)
Department of Aeronautics and Astronautics, Massachusetts Institute of
Technology, (2) Department of Electrical Engineering and Computer Science,
Massachusetts Institute of Technology, (3) Department of Aeronautics,
Imperial College London, (4) Department of Applied Mathematics, University of
Washington, (5) Department of Mechanical Engineering, University of
Washington)
- Abstract summary: We present an approach for agile and computationally efficient trajectory tracking on the MIT SoftFly, a sub-gram MAV (0.7 grams)
Our strategy employs a cascaded control scheme, where an adaptive attitude controller is combined with a neural network policy trained to imitate a trajectory tracking robust tube model predictive controller (RTMPC)
We experimentally evaluate our approach, achieving position Root Mean Square Errors lower than 1.8 cm even in the more challenging maneuvers, obtaining a 60% reduction in maximum position error compared to our previous work, and robustness demonstrating to large external disturbances.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Accurate and agile trajectory tracking in sub-gram Micro Aerial Vehicles
(MAVs) is challenging, as the small scale of the robot induces large model
uncertainties, demanding robust feedback controllers, while the fast dynamics
and computational constraints prevent the deployment of computationally
expensive strategies. In this work, we present an approach for agile and
computationally efficient trajectory tracking on the MIT SoftFly, a sub-gram
MAV (0.7 grams). Our strategy employs a cascaded control scheme, where an
adaptive attitude controller is combined with a neural network policy trained
to imitate a trajectory tracking robust tube model predictive controller
(RTMPC). The neural network policy is obtained using our recent work, which
enables the policy to preserve the robustness of RTMPC, but at a fraction of
its computational cost. We experimentally evaluate our approach, achieving
position Root Mean Square Errors lower than 1.8 cm even in the more challenging
maneuvers, obtaining a 60% reduction in maximum position error compared to our
previous work, and demonstrating robustness to large external disturbances
Related papers
- Custom Non-Linear Model Predictive Control for Obstacle Avoidance in Indoor and Outdoor Environments [0.0]
This paper introduces a Non-linear Model Predictive Control (NMPC) framework for the DJI Matrice 100.
The framework supports various trajectory types and employs a penalty-based cost function for control accuracy in tight maneuvers.
arXiv Detail & Related papers (2024-10-03T17:50:19Z) - Integrating DeepRL with Robust Low-Level Control in Robotic Manipulators for Non-Repetitive Reaching Tasks [0.24578723416255746]
In robotics, contemporary strategies are learning-based, characterized by a complex black-box nature and a lack of interpretability.
We propose integrating a collision-free trajectory planner based on deep reinforcement learning (DRL) with a novel auto-tuning low-level control strategy.
arXiv Detail & Related papers (2024-02-04T15:54:03Z) - Tube-NeRF: Efficient Imitation Learning of Visuomotor Policies from MPC
using Tube-Guided Data Augmentation and NeRFs [42.220568722735095]
Imitation learning (IL) can train computationally-efficient sensorimotor policies from a resource-intensive Model Predictive Controller (MPC)
We propose a data augmentation (DA) strategy that enables efficient learning of vision-based policies.
We show 80-fold increase in demonstration efficiency and a 50% reduction in training time over current IL methods.
arXiv Detail & Related papers (2023-11-23T18:54:25Z) - Tuning Legged Locomotion Controllers via Safe Bayesian Optimization [47.87675010450171]
This paper presents a data-driven strategy to streamline the deployment of model-based controllers in legged robotic hardware platforms.
We leverage a model-free safe learning algorithm to automate the tuning of control gains, addressing the mismatch between the simplified model used in the control formulation and the real system.
arXiv Detail & Related papers (2023-06-12T13:10:14Z) - Training Efficient Controllers via Analytic Policy Gradient [44.0762454494769]
Control design for robotic systems is complex and often requires solving an optimization to follow a trajectory accurately.
Online optimization approaches like Model Predictive Control (MPC) have been shown to achieve great tracking performance, but require high computing power.
We propose an Analytic Policy Gradient (APG) method to tackle this problem.
arXiv Detail & Related papers (2022-09-26T22:04:35Z) - Policy Search for Model Predictive Control with Application to Agile
Drone Flight [56.24908013905407]
We propose a policy-search-for-model-predictive-control framework for MPC.
Specifically, we formulate the MPC as a parameterized controller, where the hard-to-optimize decision variables are represented as high-level policies.
Experiments show that our controller achieves robust and real-time control performance in both simulation and the real world.
arXiv Detail & Related papers (2021-12-07T17:39:24Z) - Nonprehensile Riemannian Motion Predictive Control [57.295751294224765]
We introduce a novel Real-to-Sim reward analysis technique to reliably imagine and predict the outcome of taking possible actions for a real robotic platform.
We produce a closed-loop controller to reactively push objects in a continuous action space.
We observe that RMPC is robust in cluttered as well as occluded environments and outperforms the baselines.
arXiv Detail & Related papers (2021-11-15T18:50:04Z) - SABER: Data-Driven Motion Planner for Autonomously Navigating
Heterogeneous Robots [112.2491765424719]
We present an end-to-end online motion planning framework that uses a data-driven approach to navigate a heterogeneous robot team towards a global goal.
We use model predictive control (SMPC) to calculate control inputs that satisfy robot dynamics, and consider uncertainty during obstacle avoidance with chance constraints.
recurrent neural networks are used to provide a quick estimate of future state uncertainty considered in the SMPC finite-time horizon solution.
A Deep Q-learning agent is employed to serve as a high-level path planner, providing the SMPC with target positions that move the robots towards a desired global goal.
arXiv Detail & Related papers (2021-08-03T02:56:21Z) - Regret-optimal Estimation and Control [52.28457815067461]
We show that the regret-optimal estimator and regret-optimal controller can be derived in state-space form.
We propose regret-optimal analogs of Model-Predictive Control (MPC) and the Extended KalmanFilter (EKF) for systems with nonlinear dynamics.
arXiv Detail & Related papers (2021-06-22T23:14:21Z) - Towards Safe Control of Continuum Manipulator Using Shielded Multiagent
Reinforcement Learning [1.2647816797166165]
The control of the robot is formulated as a one-DoF, one agent problem in the MADQN framework to improve the learning efficiency.
Shielded MADQN enabled the robot to perform point and trajectory tracking with submillimeter root mean square errors under external loads.
arXiv Detail & Related papers (2021-06-15T05:55:05Z) - Guided Constrained Policy Optimization for Dynamic Quadrupedal Robot
Locomotion [78.46388769788405]
We introduce guided constrained policy optimization (GCPO), an RL framework based upon our implementation of constrained policy optimization (CPPO)
We show that guided constrained RL offers faster convergence close to the desired optimum resulting in an optimal, yet physically feasible, robotic control behavior without the need for precise reward function tuning.
arXiv Detail & Related papers (2020-02-22T10:15:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.