Deep Reinforcement Learning for Safe Landing Site Selection with
Concurrent Consideration of Divert Maneuvers
- URL: http://arxiv.org/abs/2102.12432v1
- Date: Wed, 24 Feb 2021 17:53:10 GMT
- Title: Deep Reinforcement Learning for Safe Landing Site Selection with
Concurrent Consideration of Divert Maneuvers
- Authors: Keidai Iiyama, Kento Tomita, Bhavi A. Jagatia, Tatsuwaki Nakagawa and
Koki Ho
- Abstract summary: This research proposes a new integrated framework for identifying safe landing locations and planning in-flight divert maneuvers.
The proposed framework was able to achieve 94.8 $%$ of successful landing in highly challenging landing sites.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This research proposes a new integrated framework for identifying safe
landing locations and planning in-flight divert maneuvers. The state-of-the-art
algorithms for landing zone selection utilize local terrain features such as
slopes and roughness to judge the safety and priority of the landing point.
However, when there are additional chances of observation and diverting in the
future, these algorithms are not able to evaluate the safety of the decision
itself to target the selected landing point considering the overall descent
trajectory. In response to this challenge, we propose a reinforcement learning
framework that optimizes a landing site selection strategy concurrently with a
guidance and control strategy to the target landing site. The trained agent
could evaluate and select landing sites with explicit consideration of the
terrain features, quality of future observations, and control to achieve a safe
and efficient landing trajectory at a system-level. The proposed framework was
able to achieve 94.8 $\%$ of successful landing in highly challenging landing
sites where over 80$\%$ of the area around the initial target lading point is
hazardous, by effectively updating the target landing site and feedback control
gain during descent.
Related papers
- Affordances-Oriented Planning using Foundation Models for Continuous Vision-Language Navigation [64.84996994779443]
We propose a novel Affordances-Oriented Planner for continuous vision-language navigation (VLN) task.
Our AO-Planner integrates various foundation models to achieve affordances-oriented low-level motion planning and high-level decision-making.
Experiments on the challenging R2R-CE and RxR-CE datasets show that AO-Planner achieves state-of-the-art zero-shot performance.
arXiv Detail & Related papers (2024-07-08T12:52:46Z) - Improving Point-based Crowd Counting and Localization Based on Auxiliary Point Guidance [59.71186244597394]
We introduce an effective approach to stabilize the proposal-target matching in point-based methods.
We propose Auxiliary Point Guidance (APG) to provide clear and effective guidance for proposal selection and optimization.
We also develop Implicit Feature Interpolation (IFI) to enable adaptive feature extraction in diverse crowd scenarios.
arXiv Detail & Related papers (2024-05-17T07:23:27Z) - Secure Navigation using Landmark-based Localization in a GPS-denied
Environment [1.19658449368018]
This paper proposes a novel framework that integrates landmark-based localization (LanBLoc) with an Extended Kalman Filter (EKF) to predict the future state of moving entities along the battlefield.
We present a simulated battlefield scenario for two different approaches that guide a moving entity through an obstacle and hazard-free path.
arXiv Detail & Related papers (2024-02-22T04:41:56Z) - Visual Environment Assessment for Safe Autonomous Quadrotor Landing [8.538463567092297]
We present a novel approach for detection and assessment of potential landing sites for safe quadrotor landing.
Our solution efficiently integrates 2D and 3D environmental information, eliminating the need for external aids such as GPS.
Our approach runs in real-time on quadrotors equipped with limited computational capabilities.
arXiv Detail & Related papers (2023-11-16T18:02:10Z) - HALO: Hazard-Aware Landing Optimization for Autonomous Systems [1.5414037351414311]
This paper presents a coupled perception-planning solution which addresses the hazard detection, optimal landing trajectory generation, and contingency planning challenges.
We develop and combine two novel algorithms, Hazard-Aware Landing Site Selection (HALSS) and Adaptive Deferred-Decision Trajectory Optimization (-DDTO), to address the perception and planning challenges.
We demonstrate the efficacy of our approach using a simulated Martian environment and show that our coupled perception-planning method achieves greater landing success.
arXiv Detail & Related papers (2023-04-04T07:20:06Z) - Meta-Learning Priors for Safe Bayesian Optimization [72.8349503901712]
We build on a meta-learning algorithm, F-PACOH, capable of providing reliable uncertainty quantification in settings of data scarcity.
As core contribution, we develop a novel framework for choosing safety-compliant priors in a data-riven manner.
On benchmark functions and a high-precision motion system, we demonstrate that our meta-learned priors accelerate the convergence of safe BO approaches.
arXiv Detail & Related papers (2022-10-03T08:38:38Z) - Detection and Initial Assessment of Lunar Landing Sites Using Neural
Networks [0.0]
This paper will focus on a passive autonomous hazard detection and avoidance sub-system to generate an initial assessment of possible landing regions for the guidance system.
The system uses a single camera and the MobileNetV2 neural network architecture to detect and discern between safe landing sites and hazards such as rocks, shadows, and craters.
arXiv Detail & Related papers (2022-07-23T04:29:18Z) - Uncertainty-driven Planner for Exploration and Navigation [36.933903274373336]
We consider the problems of exploration and point-goal navigation in previously unseen environments.
We argue that learning occupancy priors over indoor maps provides significant advantages towards addressing these problems.
We present a novel planning framework that first learns to generate occupancy maps beyond the field-of-view of the agent.
arXiv Detail & Related papers (2022-02-24T05:25:31Z) - Benchmarking Safe Deep Reinforcement Learning in Aquatic Navigation [78.17108227614928]
We propose a benchmark environment for Safe Reinforcement Learning focusing on aquatic navigation.
We consider a value-based and policy-gradient Deep Reinforcement Learning (DRL)
We also propose a verification strategy that checks the behavior of the trained models over a set of desired properties.
arXiv Detail & Related papers (2021-12-16T16:53:56Z) - Reinforcement Learning for Low-Thrust Trajectory Design of
Interplanetary Missions [77.34726150561087]
This paper investigates the use of reinforcement learning for the robust design of interplanetary trajectories in presence of severe disturbances.
An open-source implementation of the state-of-the-art algorithm Proximal Policy Optimization is adopted.
The resulting Guidance and Control Network provides both a robust nominal trajectory and the associated closed-loop guidance law.
arXiv Detail & Related papers (2020-08-19T15:22:15Z) - Chance-Constrained Trajectory Optimization for Safe Exploration and
Learning of Nonlinear Systems [81.7983463275447]
Learning-based control algorithms require data collection with abundant supervision for training.
We present a new approach for optimal motion planning with safe exploration that integrates chance-constrained optimal control with dynamics learning and feedback control.
arXiv Detail & Related papers (2020-05-09T05:57:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.