Inverted Landing in a Small Aerial Robot via Deep Reinforcement Learning
for Triggering and Control of Rotational Maneuvers
- URL: http://arxiv.org/abs/2209.11043v2
- Date: Tue, 25 Apr 2023 13:58:24 GMT
- Title: Inverted Landing in a Small Aerial Robot via Deep Reinforcement Learning
for Triggering and Control of Rotational Maneuvers
- Authors: Bryan Habas, Jack W. Langelaan, Bo Cheng
- Abstract summary: Inverted landing in a rapid and robust manner is a challenging feat for aerial robots, especially while depending entirely on onboard sensing and computation.
Previous work has identified a direct causal connection between a series of onboard visual cues and kinematic actions that allow for reliable execution of this challenging aerobatic maneuver in small aerial robots.
In this work, we first utilized Deep Reinforcement Learning and a physics-based simulation to obtain a general, optimal control policy for robust inverted landing.
- Score: 11.29285364660789
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Inverted landing in a rapid and robust manner is a challenging feat for
aerial robots, especially while depending entirely on onboard sensing and
computation. In spite of this, this feat is routinely performed by biological
fliers such as bats, flies, and bees. Our previous work has identified a direct
causal connection between a series of onboard visual cues and kinematic actions
that allow for reliable execution of this challenging aerobatic maneuver in
small aerial robots. In this work, we first utilized Deep Reinforcement
Learning and a physics-based simulation to obtain a general, optimal control
policy for robust inverted landing starting from any arbitrary approach
condition. This optimized control policy provides a computationally-efficient
mapping from the system's observational space to its motor command action
space, including both triggering and control of rotational maneuvers. This was
done by training the system over a large range of approach flight velocities
that varied with magnitude and direction.
Next, we performed a sim-to-real transfer and experimental validation of the
learned policy via domain randomization, by varying the robot's inertial
parameters in the simulation. Through experimental trials, we identified
several dominant factors which greatly improved landing robustness and the
primary mechanisms that determined inverted landing success. We expect the
learning framework developed in this study can be generalized to solve more
challenging tasks, such as utilizing noisy onboard sensory data, landing on
surfaces of various orientations, or landing on dynamically-moving surfaces.
Related papers
- Learning to enhance multi-legged robot on rugged landscapes [7.956679144631909]
Multi-legged robots offer a promising solution forNavigating rugged landscapes.
Recent studies have shown that a linear controller can ensure reliable mobility on challenging terrains.
We develop a MuJoCo-based simulator tailored to this robotic platform and use the simulation to develop a reinforcement learning-based control framework.
arXiv Detail & Related papers (2024-09-14T15:53:08Z) - Gaussian Splatting to Real World Flight Navigation Transfer with Liquid Networks [93.38375271826202]
We present a method to improve generalization and robustness to distribution shifts in sim-to-real visual quadrotor navigation tasks.
We first build a simulator by integrating Gaussian splatting with quadrotor flight dynamics, and then, train robust navigation policies using Liquid neural networks.
In this way, we obtain a full-stack imitation learning protocol that combines advances in 3D Gaussian splatting radiance field rendering, programming of expert demonstration training data, and the task understanding capabilities of Liquid networks.
arXiv Detail & Related papers (2024-06-21T13:48:37Z) - From Flies to Robots: Inverted Landing in Small Quadcopters with Dynamic
Perching [15.57055572401334]
Inverted landing is a routine behavior among a number of animal fliers.
We develop a control policy general to arbitrary ceiling-approach conditions.
We successfully achieved a range of robust inverted-landing behaviors in small quadcopters.
arXiv Detail & Related papers (2024-02-29T21:09:08Z) - Optimality Principles in Spacecraft Neural Guidance and Control [16.59877059263942]
We argue that end-to-end neural guidance and control architectures (here called G&CNets) allow transferring onboard the burden of acting upon optimality principles.
In this way, the sensor information is transformed in real time into optimal plans thus increasing the mission autonomy and robustness.
We discuss the main results obtained in training such neural architectures in simulation for interplanetary transfers, landings and close proximity operations.
arXiv Detail & Related papers (2023-05-22T14:48:58Z) - Physics-Inspired Temporal Learning of Quadrotor Dynamics for Accurate
Model Predictive Trajectory Tracking [76.27433308688592]
Accurately modeling quadrotor's system dynamics is critical for guaranteeing agile, safe, and stable navigation.
We present a novel Physics-Inspired Temporal Convolutional Network (PI-TCN) approach to learning quadrotor's system dynamics purely from robot experience.
Our approach combines the expressive power of sparse temporal convolutions and dense feed-forward connections to make accurate system predictions.
arXiv Detail & Related papers (2022-06-07T13:51:35Z) - Neural-Fly Enables Rapid Learning for Agile Flight in Strong Winds [96.74836678572582]
We present a learning-based approach that allows rapid online adaptation by incorporating pretrained representations through deep learning.
Neural-Fly achieves precise flight control with substantially smaller tracking error than state-of-the-art nonlinear and adaptive controllers.
arXiv Detail & Related papers (2022-05-13T21:55:28Z) - Learning Pneumatic Non-Prehensile Manipulation with a Mobile Blower [30.032847855193864]
blowing controller must continually adapt to unexpected changes from its actions.
We introduce a multi-frequency version of the spatial action maps framework.
This allows for efficient learning of vision-based policies that effectively combine high-level planning and low-level closed-loop control.
arXiv Detail & Related papers (2022-04-05T17:55:58Z) - Nonprehensile Riemannian Motion Predictive Control [57.295751294224765]
We introduce a novel Real-to-Sim reward analysis technique to reliably imagine and predict the outcome of taking possible actions for a real robotic platform.
We produce a closed-loop controller to reactively push objects in a continuous action space.
We observe that RMPC is robust in cluttered as well as occluded environments and outperforms the baselines.
arXiv Detail & Related papers (2021-11-15T18:50:04Z) - Learning High-Speed Flight in the Wild [101.33104268902208]
We propose an end-to-end approach that can autonomously fly quadrotors through complex natural and man-made environments at high speeds.
The key principle is to directly map noisy sensory observations to collision-free trajectories in a receding-horizon fashion.
By simulating realistic sensor noise, our approach achieves zero-shot transfer from simulation to challenging real-world environments.
arXiv Detail & Related papers (2021-10-11T09:43:11Z) - OSCAR: Data-Driven Operational Space Control for Adaptive and Robust
Robot Manipulation [50.59541802645156]
Operational Space Control (OSC) has been used as an effective task-space controller for manipulation.
We propose OSC for Adaptation and Robustness (OSCAR), a data-driven variant of OSC that compensates for modeling errors.
We evaluate our method on a variety of simulated manipulation problems, and find substantial improvements over an array of controller baselines.
arXiv Detail & Related papers (2021-10-02T01:21:38Z) - RLOC: Terrain-Aware Legged Locomotion using Reinforcement Learning and
Optimal Control [6.669503016190925]
We present a unified model-based and data-driven approach for quadrupedal planning and control.
We map sensory information and desired base velocity commands into footstep plans using a reinforcement learning policy.
We train and evaluate our framework on a complex quadrupedal system, ANYmal B, and demonstrate transferability to a larger and heavier robot, ANYmal C, without requiring retraining.
arXiv Detail & Related papers (2020-12-05T18:30:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.