TempFuser: Learning Agile, Tactical, and Acrobatic Flight Maneuvers Using a Long Short-Term Temporal Fusion Transformer
- URL: http://arxiv.org/abs/2308.03257v4
- Date: Wed, 25 Sep 2024 07:09:05 GMT
- Title: TempFuser: Learning Agile, Tactical, and Acrobatic Flight Maneuvers Using a Long Short-Term Temporal Fusion Transformer
- Authors: Hyunki Seong, David Hyunchul Shim,
- Abstract summary: TempFuser is a novel long short-term temporal fusion transformer architecture.
It can learn agile, tactical, and acrobatic flight maneuvers in complex dogfight problems.
Our model exhibits human-like acrobatic maneuvers even when facing adversaries with superior specifications.
- Score: 2.163881720692685
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Dogfighting is a challenging scenario in aerial applications that requires a comprehensive understanding of both strategic maneuvers and the aerodynamics of agile aircraft. The aerial agent needs to not only understand tactically evolving maneuvers of fighter jets from a long-term perspective but also react to rapidly changing aerodynamics of aircraft from a short-term viewpoint. In this paper, we introduce TempFuser, a novel long short-term temporal fusion transformer architecture that can learn agile, tactical, and acrobatic flight maneuvers in complex dogfight problems. Our approach integrates two distinct temporal transition embeddings into a transformer-based network to comprehensively capture both the long-term tactics and short-term agility of aerial agents. By incorporating these perspectives, our policy network generates end-to-end flight commands that secure dominant positions over the long term and effectively outmaneuver agile opponents. After training in a high-fidelity flight simulator, our model successfully learns to execute strategic maneuvers, outperforming baseline policy models against various types of opponent aircraft. Notably, our model exhibits human-like acrobatic maneuvers even when facing adversaries with superior specifications, all without relying on prior knowledge. Moreover, it demonstrates robust pursuit performance in challenging supersonic and low-altitude situations. Demo videos are available at https://sites.google.com/view/tempfuser.
Related papers
- Online Adaptation for Flying Quadrotors in Tight Formations [10.227479910430866]
Complex aerodynamic wake interactions can destabilize individual team members as well as the team.<n>We present L1 KNODE-DW MPC, an adaptive, mixed expert learning based control framework.<n>Our results show that the proposed framework is capable of enabling the three-quadrotor team to remain vertically aligned in close proximity throughout the flight.
arXiv Detail & Related papers (2025-06-20T21:49:17Z) - Training Environment for High Performance Reinforcement Learning [0.0]
Tunnel is a reinforcement learning training environment for high performance aircraft.<n>It integrates the F16 3D nonlinear flight dynamics into OpenAI Gymnasium python package.
arXiv Detail & Related papers (2025-05-04T01:09:15Z) - An Imitative Reinforcement Learning Framework for Autonomous Dogfight [20.150691753213817]
Unmanned Combat Aerial Vehicle (UCAV) dogfight plays a decisive role on the aerial battlefields.
This paper proposes a novel imitative reinforcement learning framework, which efficiently leverages expert data while enabling autonomous exploration.
The proposed framework can learn a successful dogfight policy of 'pursuit-lock-launch' for UCAVs.
arXiv Detail & Related papers (2024-06-17T13:59:52Z) - Fighter flight trajectory prediction based on spatio-temporal graphcial attention network [8.938877973527779]
This paper proposes a network-temporal graph attention (ST-GAT) using encoding and decoding structures to predict the flight trajectory.
The Transformer branch network is used to extract the characteristics of historical trajectories and capture the impact of the fighter's temporal state on future trajectories.
The GAT branch network is used to extract spatial features in historical trajectories and capture potential spatial correlations between fighters.
arXiv Detail & Related papers (2024-05-13T02:47:57Z) - From Flies to Robots: Inverted Landing in Small Quadcopters with Dynamic
Perching [15.57055572401334]
Inverted landing is a routine behavior among a number of animal fliers.
We develop a control policy general to arbitrary ceiling-approach conditions.
We successfully achieved a range of robust inverted-landing behaviors in small quadcopters.
arXiv Detail & Related papers (2024-02-29T21:09:08Z) - Reinforcement Learning for Versatile, Dynamic, and Robust Bipedal Locomotion Control [106.32794844077534]
This paper presents a study on using deep reinforcement learning to create dynamic locomotion controllers for bipedal robots.
We develop a general control solution that can be used for a range of dynamic bipedal skills, from periodic walking and running to aperiodic jumping and standing.
This work pushes the limits of agility for bipedal robots through extensive real-world experiments.
arXiv Detail & Related papers (2024-01-30T10:48:43Z) - Hierarchical Multi-Agent Reinforcement Learning for Air Combat
Maneuvering [40.06500618820166]
We propose a hierarchical multi-agent reinforcement learning framework for air-to-air combat with multiple heterogeneous agents.
Low-level policies are trained for accurate unit combat control. The commander policy is trained on mission targets given pre-trained low-level policies.
arXiv Detail & Related papers (2023-09-20T12:16:00Z) - Reinforcement Learning Based Self-play and State Stacking Techniques for
Noisy Air Combat Environment [1.7403133838762446]
The complexity of air combat arises from aggressive close-range maneuvers and agile enemy behaviors.
In this study, we developed an air combat simulation, which provides noisy observations to the agents.
We present a state stacking method for noisy RL environments as a noise reduction technique.
arXiv Detail & Related papers (2023-03-06T12:23:23Z) - Robust and Versatile Bipedal Jumping Control through Reinforcement
Learning [141.56016556936865]
This work aims to push the limits of agility for bipedal robots by enabling a torque-controlled bipedal robot to perform robust and versatile dynamic jumps in the real world.
We present a reinforcement learning framework for training a robot to accomplish a large variety of jumping tasks, such as jumping to different locations and directions.
We develop a new policy structure that encodes the robot's long-term input/output (I/O) history while also providing direct access to a short-term I/O history.
arXiv Detail & Related papers (2023-02-19T01:06:09Z) - Neural-Fly Enables Rapid Learning for Agile Flight in Strong Winds [96.74836678572582]
We present a learning-based approach that allows rapid online adaptation by incorporating pretrained representations through deep learning.
Neural-Fly achieves precise flight control with substantially smaller tracking error than state-of-the-art nonlinear and adaptive controllers.
arXiv Detail & Related papers (2022-05-13T21:55:28Z) - Fixed Points in Cyber Space: Rethinking Optimal Evasion Attacks in the
Age of AI-NIDS [70.60975663021952]
We study blackbox adversarial attacks on network classifiers.
We argue that attacker-defender fixed points are themselves general-sum games with complex phase transitions.
We show that a continual learning approach is required to study attacker-defender dynamics.
arXiv Detail & Related papers (2021-11-23T23:42:16Z) - Learning Agile Locomotion via Adversarial Training [59.03007947334165]
In this paper, we present a multi-agent learning system, in which a quadruped robot (protagonist) learns to chase another robot (adversary) while the latter learns to escape.
We find that this adversarial training process not only encourages agile behaviors but also effectively alleviates the laborious environment design effort.
In contrast to prior works that used only one adversary, we find that training an ensemble of adversaries, each of which specializes in a different escaping strategy, is essential for the protagonist to master agility.
arXiv Detail & Related papers (2020-08-03T01:20:37Z) - Model-Based Meta-Reinforcement Learning for Flight with Suspended
Payloads [69.21503033239985]
Transporting suspended payloads is challenging for autonomous aerial vehicles.
We propose a meta-learning approach that "learns how to learn" models of altered dynamics within seconds of post-connection flight data.
arXiv Detail & Related papers (2020-04-23T17:43:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.