Real-Time Optimal Guidance and Control for Interplanetary Transfers
Using Deep Networks
- URL: http://arxiv.org/abs/2002.09063v1
- Date: Thu, 20 Feb 2020 23:37:43 GMT
- Title: Real-Time Optimal Guidance and Control for Interplanetary Transfers
Using Deep Networks
- Authors: Dario Izzo and Ekin \"Ozt\"urk
- Abstract summary: Imitation learning of optimal examples is used as a network training paradigm.
G&CNETs are suitable for an on-board, real-time, implementation of the optimal guidance and control system of the spacecraft.
- Score: 10.191757341020216
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We consider the Earth-Venus mass-optimal interplanetary transfer of a
low-thrust spacecraft and show how the optimal guidance can be represented by
deep networks in a large portion of the state space and to a high degree of
accuracy. Imitation (supervised) learning of optimal examples is used as a
network training paradigm. The resulting models are suitable for an on-board,
real-time, implementation of the optimal guidance and control system of the
spacecraft and are called G&CNETs. A new general methodology called Backward
Generation of Optimal Examples is introduced and shown to be able to
efficiently create all the optimal state action pairs necessary to train
G&CNETs without solving optimal control problems. With respect to previous
works, we are able to produce datasets containing a few orders of magnitude
more optimal trajectories and obtain network performances compatible with real
missions requirements. Several schemes able to train representations of either
the optimal policy (thrust profile) or the value function (optimal mass) are
proposed and tested. We find that both policy learning and value function
learning successfully and accurately learn the optimal thrust and that a
spacecraft employing the learned thrust is able to reach the target conditions
orbit spending only 2 permil more propellant than in the corresponding
mathematically optimal transfer. Moreover, the optimal propellant mass can be
predicted (in case of value function learning) within an error well within 1%.
All G&CNETs produced are tested during simulations of interplanetary transfers
with respect to their ability to reach the target conditions optimally starting
from nominal and off-nominal conditions.
Related papers
- Towards Robust Spacecraft Trajectory Optimization via Transformers [17.073280827888226]
Future multi-spacecraft missions require robust autonomous optimization capabilities to ensure safe and efficient rendezvous operations.
To mitigate this burden, introduced generative Transformer model to provide robust optimal initial guesses.
This work extends capabilities of ART to address robustconstrained optimal control problems.
arXiv Detail & Related papers (2024-10-08T00:58:42Z) - Hallmarks of Optimization Trajectories in Neural Networks: Directional Exploration and Redundancy [75.15685966213832]
We analyze the rich directional structure of optimization trajectories represented by their pointwise parameters.
We show that training only scalar batchnorm parameters some while into training matches the performance of training the entire network.
arXiv Detail & Related papers (2024-03-12T07:32:47Z) - Learning Regions of Interest for Bayesian Optimization with Adaptive
Level-Set Estimation [84.0621253654014]
We propose a framework, called BALLET, which adaptively filters for a high-confidence region of interest.
We show theoretically that BALLET can efficiently shrink the search space, and can exhibit a tighter regret bound than standard BO.
arXiv Detail & Related papers (2023-07-25T09:45:47Z) - Optimality Principles in Spacecraft Neural Guidance and Control [16.59877059263942]
We argue that end-to-end neural guidance and control architectures (here called G&CNets) allow transferring onboard the burden of acting upon optimality principles.
In this way, the sensor information is transformed in real time into optimal plans thus increasing the mission autonomy and robustness.
We discuss the main results obtained in training such neural architectures in simulation for interplanetary transfers, landings and close proximity operations.
arXiv Detail & Related papers (2023-05-22T14:48:58Z) - Gradient-Based Trajectory Optimization With Learned Dynamics [80.41791191022139]
We use machine learning techniques to learn a differentiable dynamics model of the system from data.
We show that a neural network can model highly nonlinear behaviors accurately for large time horizons.
In our hardware experiments, we demonstrate that our learned model can represent complex dynamics for both the Spot and Radio-controlled (RC) car.
arXiv Detail & Related papers (2022-04-09T22:07:34Z) - Neural representation of a time optimal, constant acceleration
rendezvous [10.191757341020216]
We train neural models to represent both the optimal policy (i.e. the optimal thrust direction) and the value function (i.e. the time of flight) for a time optimal, constant acceleration low-thrust rendezvous.
We achieve, in all cases, accuracies resulting in successful rendezvous and time of flight predictions.
arXiv Detail & Related papers (2022-03-29T12:40:50Z) - Bayesian Optimization and Deep Learning forsteering wheel angle
prediction [58.720142291102135]
This work aims to obtain an accurate model for the prediction of the steering angle in an automated driving system.
BO was able to identify, within a limited number of trials, a model -- namely BOST-LSTM -- which resulted, the most accurate when compared to classical end-to-end driving models.
arXiv Detail & Related papers (2021-10-22T15:25:14Z) - Reinforcement Learning for Low-Thrust Trajectory Design of
Interplanetary Missions [77.34726150561087]
This paper investigates the use of reinforcement learning for the robust design of interplanetary trajectories in presence of severe disturbances.
An open-source implementation of the state-of-the-art algorithm Proximal Policy Optimization is adopted.
The resulting Guidance and Control Network provides both a robust nominal trajectory and the associated closed-loop guidance law.
arXiv Detail & Related papers (2020-08-19T15:22:15Z) - Localized active learning of Gaussian process state space models [63.97366815968177]
A globally accurate model is not required to achieve good performance in many common control applications.
We propose an active learning strategy for Gaussian process state space models that aims to obtain an accurate model on a bounded subset of the state-action space.
By employing model predictive control, the proposed technique integrates information collected during exploration and adaptively improves its exploration strategy.
arXiv Detail & Related papers (2020-05-04T05:35:02Z) - Reinforcement Meta-Learning for Interception of Maneuvering
Exoatmospheric Targets with Parasitic Attitude Loop [1.7663909228482466]
We use Reinforcement Meta-Learning to optimize an adaptive integrated guidance, navigation, and control system suitable for exoatmospheric interception of a maneuvering target.
The system maps observations consisting of strapdown seeker angles and rate gyro measurements directly to thruster on-off commands.
We demonstrate that the optimized policy can adapt to parasitic effects including seeker angle measurement lag, thruster control lag, the parasitic attitude loop resulting from scale factor errors and Gaussian noise on angle and rotational velocity measurements.
arXiv Detail & Related papers (2020-04-18T21:20:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.