Spacecraft Autonomous Decision-Planning for Collision Avoidance: a
Reinforcement Learning Approach
- URL: http://arxiv.org/abs/2310.18966v1
- Date: Sun, 29 Oct 2023 10:15:33 GMT
- Title: Spacecraft Autonomous Decision-Planning for Collision Avoidance: a
Reinforcement Learning Approach
- Authors: Nicolas Bourriez, Adrien Loizeau and Adam F. Abdin
- Abstract summary: This work proposes an implementation of autonomous CA decision-making capabilities on spacecraft based on Reinforcement Learning techniques.
The proposed framework considers imperfect monitoring information about the status of the debris in orbit and allows the AI system to effectively learn policies to perform accurate Collision Avoidance Maneuvers (CAMs)
The objective is to successfully delegate the decision-making process for autonomously implementing a CAM to the spacecraft without human intervention.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The space environment around the Earth is becoming increasingly populated by
both active spacecraft and space debris. To avoid potential collision events,
significant improvements in Space Situational Awareness (SSA) activities and
Collision Avoidance (CA) technologies are allowing the tracking and maneuvering
of spacecraft with increasing accuracy and reliability. However, these
procedures still largely involve a high level of human intervention to make the
necessary decisions. For an increasingly complex space environment, this
decision-making strategy is not likely to be sustainable. Therefore, it is
important to successfully introduce higher levels of automation for key Space
Traffic Management (STM) processes to ensure the level of reliability needed
for navigating a large number of spacecraft. These processes range from
collision risk detection to the identification of the appropriate action to
take and the execution of avoidance maneuvers. This work proposes an
implementation of autonomous CA decision-making capabilities on spacecraft
based on Reinforcement Learning (RL) techniques. A novel methodology based on a
Partially Observable Markov Decision Process (POMDP) framework is developed to
train the Artificial Intelligence (AI) system on board the spacecraft,
considering epistemic and aleatory uncertainties. The proposed framework
considers imperfect monitoring information about the status of the debris in
orbit and allows the AI system to effectively learn stochastic policies to
perform accurate Collision Avoidance Maneuvers (CAMs). The objective is to
successfully delegate the decision-making process for autonomously implementing
a CAM to the spacecraft without human intervention. This approach would allow
for a faster response in the decision-making process and for highly
decentralized operations.
Related papers
- On-orbit Servicing for Spacecraft Collision Avoidance With Autonomous Decision Making [0.0]
This study develops an AI-based implementation of autonomous On-Orbit Servicing (OOS) mission to assist with spacecraft collision avoidance maneuvers (CAMs)
We propose an autonomous servicer' trained with Reinforcement Learning (RL) to autonomously detect potential collisions between a target satellite and space debris, rendezvous and dock with endangered satellites, and execute optimal CAM.
arXiv Detail & Related papers (2024-09-25T17:40:37Z) - AI-Driven Risk-Aware Scheduling for Active Debris Removal Missions [0.0]
Debris in Low Earth Orbit represents a significant threat to space sustainability and spacecraft safety.
Armoured Transfer Vehicles (OTVs) facilitate debris deorbiting, thereby reducing future collision risks.
Armoured decision-planning model based on Deep Reinforcement Learning (DRL) is developed to train an OTV to plan optimal debris removal sequencing.
It is shown that using the proposed framework, the agent can find optimal mission plans and learn to update the planning autonomously to include risk handling of debris with high collision risk.
arXiv Detail & Related papers (2024-09-25T15:16:07Z) - Aquatic Navigation: A Challenging Benchmark for Deep Reinforcement Learning [53.3760591018817]
We propose a new benchmarking environment for aquatic navigation using recent advances in the integration between game engines and Deep Reinforcement Learning.
Specifically, we focus on PPO, one of the most widely accepted algorithms, and we propose advanced training techniques.
Our empirical evaluation shows that a well-designed combination of these ingredients can achieve promising results.
arXiv Detail & Related papers (2024-05-30T23:20:23Z) - RACER: Epistemic Risk-Sensitive RL Enables Fast Driving with Fewer Crashes [57.319845580050924]
We propose a reinforcement learning framework that combines risk-sensitive control with an adaptive action space curriculum.
We show that our algorithm is capable of learning high-speed policies for a real-world off-road driving task.
arXiv Detail & Related papers (2024-05-07T23:32:36Z) - Convergence of Communications, Control, and Machine Learning for Secure
and Autonomous Vehicle Navigation [78.60496411542549]
Connected and autonomous vehicles (CAVs) can reduce human errors in traffic accidents, increase road efficiency, and execute various tasks. Reaping these benefits requires CAVs to autonomously navigate to target destinations.
This article proposes solutions using the convergence of communication theory, control theory, and machine learning to enable effective and secure CAV navigation.
arXiv Detail & Related papers (2023-07-05T21:38:36Z) - HALO: Hazard-Aware Landing Optimization for Autonomous Systems [1.5414037351414311]
This paper presents a coupled perception-planning solution which addresses the hazard detection, optimal landing trajectory generation, and contingency planning challenges.
We develop and combine two novel algorithms, Hazard-Aware Landing Site Selection (HALSS) and Adaptive Deferred-Decision Trajectory Optimization (-DDTO), to address the perception and planning challenges.
We demonstrate the efficacy of our approach using a simulated Martian environment and show that our coupled perception-planning method achieves greater landing success.
arXiv Detail & Related papers (2023-04-04T07:20:06Z) - Reinforcement Learning-Based Air Traffic Deconfliction [7.782300855058585]
This work focuses on automating the horizontal separation of two aircraft and presents the obstacle avoidance problem as a 2D surrogate optimization task.
Using Reinforcement Learning (RL), we optimize the avoidance policy and model the dynamics, interactions, and decision-making.
The proposed system generates a quick and achievable avoidance trajectory that satisfies the safety requirements.
arXiv Detail & Related papers (2023-01-05T00:37:20Z) - Risk-based implementation of COLREGs for autonomous surface vehicles
using deep reinforcement learning [1.304892050913381]
Deep reinforcement learning (DRL) has shown great potential for a wide range of applications.
In this work, a subset of the International Regulations for Preventing Collisions at Sea (COLREGs) is incorporated into a DRL-based path following and obstacle avoidance system.
The resulting autonomous agent dynamically interpolates between path following and COLREG-compliant collision avoidance in the training scenario, isolated encounter situations, and AIS-based simulations of real-world scenarios.
arXiv Detail & Related papers (2021-11-30T21:32:59Z) - OSCAR: Data-Driven Operational Space Control for Adaptive and Robust
Robot Manipulation [50.59541802645156]
Operational Space Control (OSC) has been used as an effective task-space controller for manipulation.
We propose OSC for Adaptation and Robustness (OSCAR), a data-driven variant of OSC that compensates for modeling errors.
We evaluate our method on a variety of simulated manipulation problems, and find substantial improvements over an array of controller baselines.
arXiv Detail & Related papers (2021-10-02T01:21:38Z) - SABER: Data-Driven Motion Planner for Autonomously Navigating
Heterogeneous Robots [112.2491765424719]
We present an end-to-end online motion planning framework that uses a data-driven approach to navigate a heterogeneous robot team towards a global goal.
We use model predictive control (SMPC) to calculate control inputs that satisfy robot dynamics, and consider uncertainty during obstacle avoidance with chance constraints.
recurrent neural networks are used to provide a quick estimate of future state uncertainty considered in the SMPC finite-time horizon solution.
A Deep Q-learning agent is employed to serve as a high-level path planner, providing the SMPC with target positions that move the robots towards a desired global goal.
arXiv Detail & Related papers (2021-08-03T02:56:21Z) - Reinforcement Learning for Low-Thrust Trajectory Design of
Interplanetary Missions [77.34726150561087]
This paper investigates the use of reinforcement learning for the robust design of interplanetary trajectories in presence of severe disturbances.
An open-source implementation of the state-of-the-art algorithm Proximal Policy Optimization is adopted.
The resulting Guidance and Control Network provides both a robust nominal trajectory and the associated closed-loop guidance law.
arXiv Detail & Related papers (2020-08-19T15:22:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.