Learning-to-Fly: Learning-based Collision Avoidance for Scalable Urban
Air Mobility
- URL: http://arxiv.org/abs/2006.13267v1
- Date: Tue, 23 Jun 2020 18:46:31 GMT
- Title: Learning-to-Fly: Learning-based Collision Avoidance for Scalable Urban
Air Mobility
- Authors: Al\"ena Rodionova, Yash Vardhan Pant, Kuk Jang, Houssam Abbas and
Rahul Mangharam
- Abstract summary: We present Learning-to-Fly (L2F), a decentralized on-demand airborne collision avoidance framework for multiple UAS.
L2F is a two-stage collision avoidance method that consists of: 1) a learning-based decision-making scheme and 2) a distributed, linear programming-based UAS control algorithm.
We show the real-time applicability of our method which is $approx!6000times$ faster than the MILP approach and can resolve $100%$ of collisions when there is ample room to maneuver.
- Score: 2.117421588033177
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: With increasing urban population, there is global interest in Urban Air
Mobility (UAM), where hundreds of autonomous Unmanned Aircraft Systems (UAS)
execute missions in the airspace above cities. Unlike traditional
human-in-the-loop air traffic management, UAM requires decentralized autonomous
approaches that scale for an order of magnitude higher aircraft densities and
are applicable to urban settings. We present Learning-to-Fly (L2F), a
decentralized on-demand airborne collision avoidance framework for multiple UAS
that allows them to independently plan and safely execute missions with
spatial, temporal and reactive objectives expressed using Signal Temporal
Logic. We formulate the problem of predictively avoiding collisions between two
UAS without violating mission objectives as a Mixed Integer Linear Program
(MILP).This however is intractable to solve online. Instead, we develop L2F, a
two-stage collision avoidance method that consists of: 1) a learning-based
decision-making scheme and 2) a distributed, linear programming-based UAS
control algorithm. Through extensive simulations, we show the real-time
applicability of our method which is $\approx\!6000\times$ faster than the MILP
approach and can resolve $100\%$ of collisions when there is ample room to
maneuver, and shows graceful degradation in performance otherwise. We also
compare L2F to two other methods and demonstrate an implementation on
quad-rotor robots.
Related papers
- Generalizability of Graph Neural Networks for Decentralized Unlabeled Motion Planning [72.86540018081531]
Unlabeled motion planning involves assigning a set of robots to target locations while ensuring collision avoidance.
This problem forms an essential building block for multi-robot systems in applications such as exploration, surveillance, and transportation.
We address this problem in a decentralized setting where each robot knows only the positions of its $k$-nearest robots and $k$-nearest targets.
arXiv Detail & Related papers (2024-09-29T23:57:25Z) - Efficient Adversarial Training in LLMs with Continuous Attacks [99.5882845458567]
Large language models (LLMs) are vulnerable to adversarial attacks that can bypass their safety guardrails.
We propose a fast adversarial training algorithm (C-AdvUL) composed of two losses.
C-AdvIPO is an adversarial variant of IPO that does not require utility data for adversarially robust alignment.
arXiv Detail & Related papers (2024-05-24T14:20:09Z) - Toward collision-free trajectory for autonomous and pilot-controlled
unmanned aerial vehicles [1.018017727755629]
This study makes greater use of electronic conspicuity (EC) information made available by PilotAware Ltd in developing an advanced collision management methodology.
The merits of the DACM methodology have been demonstrated through extensive simulations and real-world field tests in avoiding mid-air collisions.
arXiv Detail & Related papers (2023-09-18T18:24:31Z) - Reinforcement Learning-Based Air Traffic Deconfliction [7.782300855058585]
This work focuses on automating the horizontal separation of two aircraft and presents the obstacle avoidance problem as a 2D surrogate optimization task.
Using Reinforcement Learning (RL), we optimize the avoidance policy and model the dynamics, interactions, and decision-making.
The proposed system generates a quick and achievable avoidance trajectory that satisfies the safety requirements.
arXiv Detail & Related papers (2023-01-05T00:37:20Z) - NeurIPS 2022 Competition: Driving SMARTS [60.948652154552136]
Driving SMARTS is a regular competition designed to tackle problems caused by the distribution shift in dynamic interaction contexts.
The proposed competition supports methodologically diverse solutions, such as reinforcement learning (RL) and offline learning methods.
arXiv Detail & Related papers (2022-11-14T17:10:53Z) - Risk-based implementation of COLREGs for autonomous surface vehicles
using deep reinforcement learning [1.304892050913381]
Deep reinforcement learning (DRL) has shown great potential for a wide range of applications.
In this work, a subset of the International Regulations for Preventing Collisions at Sea (COLREGs) is incorporated into a DRL-based path following and obstacle avoidance system.
The resulting autonomous agent dynamically interpolates between path following and COLREG-compliant collision avoidance in the training scenario, isolated encounter situations, and AIS-based simulations of real-world scenarios.
arXiv Detail & Related papers (2021-11-30T21:32:59Z) - Obstacle Avoidance for UAS in Continuous Action Space Using Deep
Reinforcement Learning [9.891207216312937]
Obstacle avoidance for small unmanned aircraft is vital for the safety of future urban air mobility.
We propose a deep reinforcement learning algorithm based on Proximal Policy Optimization (PPO) to guide autonomous UAS to their destinations.
Results show that the proposed model can provide accurate and robust guidance and resolve conflict with a success rate of over 99%.
arXiv Detail & Related papers (2021-11-13T04:44:53Z) - SABER: Data-Driven Motion Planner for Autonomously Navigating
Heterogeneous Robots [112.2491765424719]
We present an end-to-end online motion planning framework that uses a data-driven approach to navigate a heterogeneous robot team towards a global goal.
We use model predictive control (SMPC) to calculate control inputs that satisfy robot dynamics, and consider uncertainty during obstacle avoidance with chance constraints.
recurrent neural networks are used to provide a quick estimate of future state uncertainty considered in the SMPC finite-time horizon solution.
A Deep Q-learning agent is employed to serve as a high-level path planner, providing the SMPC with target positions that move the robots towards a desired global goal.
arXiv Detail & Related papers (2021-08-03T02:56:21Z) - Learning-'N-Flying: A Learning-based, Decentralized Mission Aware UAS
Collision Avoidance Scheme [0.0]
Learning-'N-Flying (LNF) is a multi-UAS Collision Avoidance (CA) framework.
It is decentralized, works on-the-fly and allows autonomous UAS managed by different operators to safely carry out complex missions.
We show that our method can run online (computation time in the order of milliseconds), and under certain assumptions has failure rates of less than 1% in the worst-case.
arXiv Detail & Related papers (2021-01-25T20:38:17Z) - ReLMoGen: Leveraging Motion Generation in Reinforcement Learning for
Mobile Manipulation [99.2543521972137]
ReLMoGen is a framework that combines a learned policy to predict subgoals and a motion generator to plan and execute the motion needed to reach these subgoals.
Our method is benchmarked on a diverse set of seven robotics tasks in photo-realistic simulation environments.
ReLMoGen shows outstanding transferability between different motion generators at test time, indicating a great potential to transfer to real robots.
arXiv Detail & Related papers (2020-08-18T08:05:15Z) - Congestion-aware Evacuation Routing using Augmented Reality Devices [96.68280427555808]
We present a congestion-aware routing solution for indoor evacuation, which produces real-time individual-customized evacuation routes among multiple destinations.
A population density map, obtained on-the-fly by aggregating locations of evacuees from user-end Augmented Reality (AR) devices, is used to model the congestion distribution inside a building.
arXiv Detail & Related papers (2020-04-25T22:54:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.