Large Language Models as Autonomous Spacecraft Operators in Kerbal Space Program
- URL: http://arxiv.org/abs/2505.19896v1
- Date: Mon, 26 May 2025 12:25:35 GMT
- Title: Large Language Models as Autonomous Spacecraft Operators in Kerbal Space Program
- Authors: Alejandro Carrasco, Victor Rodriguez-Fernandez, Richard Linares,
- Abstract summary: Large Language Models (LLMs) are autonomous agents that take actions based on the content of the user text prompts.<n>We have developed a pure LLM-based solution for the Kerbal Space Program Differential Games (KSPDG) challenge.
- Score: 42.87968485876435
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Recent trends are emerging in the use of Large Language Models (LLMs) as autonomous agents that take actions based on the content of the user text prompts. We intend to apply these concepts to the field of Control in space, enabling LLMs to play a significant role in the decision-making process for autonomous satellite operations. As a first step towards this goal, we have developed a pure LLM-based solution for the Kerbal Space Program Differential Games (KSPDG) challenge, a public software design competition where participants create autonomous agents for maneuvering satellites involved in non-cooperative space operations, running on the KSP game engine. Our approach leverages prompt engineering, few-shot prompting, and fine-tuning techniques to create an effective LLM-based agent that ranked 2nd in the competition. To the best of our knowledge, this work pioneers the integration of LLM agents into space research. The project comprises several open repositories to facilitate replication and further research. The codebase is accessible on \href{https://github.com/ARCLab-MIT/kspdg}{GitHub}, while the trained models and datasets are available on \href{https://huggingface.co/OhhTuRnz}{Hugging Face}. Additionally, experiment tracking and detailed results can be reviewed on \href{https://wandb.ai/carrusk/huggingface}{Weights \& Biases
Related papers
- Deploying Foundation Model-Enabled Air and Ground Robots in the Field: Challenges and Opportunities [65.98704516122228]
The integration of foundation models (FMs) into robotics has enabled robots to understand natural language and reason about the semantics in their environments.<n>This paper addresses the deployment of FM-enabled robots in the field, where missions often require a robot to operate in large-scale and unstructured environments.<n>We present the first demonstration of large-scale LLM-enabled robot planning in unstructured environments with several kilometers of missions.
arXiv Detail & Related papers (2025-05-14T15:28:43Z) - SpaceR: Reinforcing MLLMs in Video Spatial Reasoning [70.7401015322983]
Video spatial reasoning poses a significant challenge for existing Multimodal Large Language Models (MLLMs)<n>This limitation stems primarily from 1) the absence of high-quality datasets for this task, and 2) the lack of effective training strategies to develop spatial reasoning capabilities.<n>Motivated by the success of Reinforcement Learning with Verifiable Reward (RLVR) in unlocking spatial reasoning abilities, this aims to improve MLLMs in video spatial reasoning through the RLVR paradigm.
arXiv Detail & Related papers (2025-04-02T15:12:17Z) - Visual Language Models as Operator Agents in the Space Domain [36.943670587532026]
Vision-Language Models (VLMs) can enhance autonomous control and decision-making in space missions.<n>In the software context, we employ VLMs to interpret visual screenshots of the graphical user interface to perform complex orbital maneuvers.<n>In the hardware context, we integrate VLMs with robotic systems equipped with cameras to inspect and diagnose physical space objects, such as satellites.
arXiv Detail & Related papers (2025-01-14T03:03:37Z) - Sparse Rewards Can Self-Train Dialogue Agents [22.799506097310008]
We introduce a novel self-improvement paradigm that empowers LLM agents to autonomously enhance their performance without external human feedback.
We present ToolWOZ, a sparse reward tool-calling simulation environment derived from MultiWOZ.
We demonstrate that models trained with JOSH, both small and frontier, significantly improve tool-based interactions while preserving general model capabilities across diverse benchmarks.
arXiv Detail & Related papers (2024-09-06T21:00:57Z) - Fine-tuning LLMs for Autonomous Spacecraft Control: A Case Study Using Kerbal Space Program [42.87968485876435]
This study explores the use of fine-tuned Large Language Models (LLMs) for autonomous spacecraft control.
We demonstrate how these models can effectively control spacecraft using language-based inputs and outputs.
arXiv Detail & Related papers (2024-08-16T11:43:31Z) - LLMSat: A Large Language Model-Based Goal-Oriented Agent for Autonomous Space Exploration [0.0]
This work explores the application of Large Language Models (LLMs) as the high-level control system of a spacecraft.
A series of deep space mission scenarios simulated within the popular game engine Kerbal Space Program are used as case studies to evaluate the implementation against the requirements.
arXiv Detail & Related papers (2024-04-13T03:33:17Z) - Language Models are Spacecraft Operators [36.943670587532026]
Large Language Models (LLMs) are autonomous agents that take actions based on the content of the user text prompts.
We have developed a pure LLM-based solution for the Kerbal Space Program Differential Games (KSPDG) challenge.
arXiv Detail & Related papers (2024-03-30T16:43:59Z) - ReLMoGen: Leveraging Motion Generation in Reinforcement Learning for
Mobile Manipulation [99.2543521972137]
ReLMoGen is a framework that combines a learned policy to predict subgoals and a motion generator to plan and execute the motion needed to reach these subgoals.
Our method is benchmarked on a diverse set of seven robotics tasks in photo-realistic simulation environments.
ReLMoGen shows outstanding transferability between different motion generators at test time, indicating a great potential to transfer to real robots.
arXiv Detail & Related papers (2020-08-18T08:05:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.