Harfang3D Dog-Fight Sandbox: A Reinforcement Learning Research Platform
for the Customized Control Tasks of Fighter Aircrafts
- URL: http://arxiv.org/abs/2210.07282v1
- Date: Thu, 13 Oct 2022 18:18:09 GMT
- Title: Harfang3D Dog-Fight Sandbox: A Reinforcement Learning Research Platform
for the Customized Control Tasks of Fighter Aircrafts
- Authors: Muhammed Murat \"Ozbek and S\"uleyman Y{\i}ld{\i}r{\i}m and Muhammet
Aksoy and Eric Kernin and Emre Koyuncu
- Abstract summary: We present a semi-realistic flight simulation environment Harfang3D Dog-Fight Sandbox for fighter aircrafts.
It is aimed to be a flexible toolbox for the investigation of main challenges in aviation studies using Reinforcement Learning.
Software also allows deployment of bot aircrafts and development of multi-agent tasks.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: The advent of deep learning (DL) gave rise to significant breakthroughs in
Reinforcement Learning (RL) research. Deep Reinforcement Learning (DRL)
algorithms have reached super-human level skills when applied to vision-based
control problems as such in Atari 2600 games where environment states were
extracted from pixel information. Unfortunately, these environments are far
from being applicable to highly dynamic and complex real-world tasks as in
autonomous control of a fighter aircraft since these environments only involve
2D representation of a visual world. Here, we present a semi-realistic flight
simulation environment Harfang3D Dog-Fight Sandbox for fighter aircrafts. It is
aimed to be a flexible toolbox for the investigation of main challenges in
aviation studies using Reinforcement Learning. The program provides easy access
to flight dynamics model, environment states, and aerodynamics of the plane
enabling user to customize any specific task in order to build intelligent
decision making (control) systems via RL. The software also allows deployment
of bot aircrafts and development of multi-agent tasks. This way, multiple
groups of aircrafts can be configured to be competitive or cooperative agents
to perform complicated tasks including Dog Fight. During the experiments, we
carried out training for two different scenarios: navigating to a designated
location and within visual range (WVR) combat, shortly Dog Fight. Using Deep
Reinforcement Learning techniques for both scenarios, we were able to train
competent agents that exhibit human-like behaviours. Based on this results, it
is confirmed that Harfang3D Dog-Fight Sandbox can be utilized as a 3D realistic
RL research platform.
Related papers
- Gaussian Splatting to Real World Flight Navigation Transfer with Liquid Networks [93.38375271826202]
We present a method to improve generalization and robustness to distribution shifts in sim-to-real visual quadrotor navigation tasks.
We first build a simulator by integrating Gaussian splatting with quadrotor flight dynamics, and then, train robust navigation policies using Liquid neural networks.
In this way, we obtain a full-stack imitation learning protocol that combines advances in 3D Gaussian splatting radiance field rendering, programming of expert demonstration training data, and the task understanding capabilities of Liquid networks.
arXiv Detail & Related papers (2024-06-21T13:48:37Z) - An Imitative Reinforcement Learning Framework for Autonomous Dogfight [20.150691753213817]
Unmanned Combat Aerial Vehicle (UCAV) dogfight plays a decisive role on the aerial battlefields.
This paper proposes a novel imitative reinforcement learning framework, which efficiently leverages expert data while enabling autonomous exploration.
The proposed framework can learn a successful dogfight policy of 'pursuit-lock-launch' for UCAVs.
arXiv Detail & Related papers (2024-06-17T13:59:52Z) - Reinforcement Learning for Versatile, Dynamic, and Robust Bipedal Locomotion Control [106.32794844077534]
This paper presents a study on using deep reinforcement learning to create dynamic locomotion controllers for bipedal robots.
We develop a general control solution that can be used for a range of dynamic bipedal skills, from periodic walking and running to aperiodic jumping and standing.
This work pushes the limits of agility for bipedal robots through extensive real-world experiments.
arXiv Detail & Related papers (2024-01-30T10:48:43Z) - DIAMBRA Arena: a New Reinforcement Learning Platform for Research and
Experimentation [91.3755431537592]
This work presents DIAMBRA Arena, a new platform for reinforcement learning research and experimentation.
It features a collection of high-quality environments exposing a Python API fully compliant with OpenAI Gym standard.
They are episodic tasks with discrete actions and observations composed by raw pixels plus additional numerical values.
arXiv Detail & Related papers (2022-10-19T14:39:10Z) - WILD-SCAV: Benchmarking FPS Gaming AI on Unity3D-based Environments [5.020816812380825]
Recent advances in deep reinforcement learning (RL) have demonstrated complex decision-making capabilities in simulation environments.
However, they are hardly to more complicated problems, due to the lack of complexity and variations in the environments they are trained and tested on.
We developed WILD-SCAV, a powerful and open-world environment based on a 3D open-world FPS game to bridge the gap.
It provides realistic 3D environments of variable complexity, various tasks, and multiple modes of interaction, where agents can learn to perceive 3D environments, navigate and plan, compete and cooperate in a human-like manner
arXiv Detail & Related papers (2022-10-14T13:39:41Z) - Parallel Reinforcement Learning Simulation for Visual Quadrotor
Navigation [4.597465975849579]
Reinforcement learning (RL) is an agent-based approach for teaching robots to navigate within the physical world.
We present a simulation framework, built on AirSim, which provides efficient parallel training.
Building on this framework, Ape-X is modified to incorporate decentralised training of AirSim environments.
arXiv Detail & Related papers (2022-09-22T15:27:42Z) - Unsupervised Learning of Efficient Geometry-Aware Neural Articulated
Representations [89.1388369229542]
We propose an unsupervised method for 3D geometry-aware representation learning of articulated objects.
We obviate this need by learning the representations with GAN training.
Experiments demonstrate the efficiency of our method and show that GAN-based training enables learning of controllable 3D representations without supervision.
arXiv Detail & Related papers (2022-04-19T12:10:18Z) - Learning to Fly -- a Gym Environment with PyBullet Physics for
Reinforcement Learning of Multi-agent Quadcopter Control [0.0]
We propose an open-source environment for multiple quadcopters based on the Bullet physics engine.
Its multi-agent and vision based reinforcement learning interfaces, as well as the support of realistic collisions and aerodynamic effects, make it, to the best of our knowledge, a first of its kind.
arXiv Detail & Related papers (2021-03-03T02:47:59Z) - Robust Reinforcement Learning-based Autonomous Driving Agent for
Simulation and Real World [0.0]
We present a DRL-based algorithm that is capable of performing autonomous robot control using Deep Q-Networks (DQN)
In our approach, the agent is trained in a simulated environment and it is able to navigate both in a simulated and real-world environment.
The trained agent is able to run on limited hardware resources and its performance is comparable to state-of-the-art approaches.
arXiv Detail & Related papers (2020-09-23T15:23:54Z) - Meta-Reinforcement Learning for Robotic Industrial Insertion Tasks [70.56451186797436]
We study how to use meta-reinforcement learning to solve the bulk of the problem in simulation.
We demonstrate our approach by training an agent to successfully perform challenging real-world insertion tasks.
arXiv Detail & Related papers (2020-04-29T18:00:22Z) - Improving Target-driven Visual Navigation with Attention on 3D Spatial
Relationships [52.72020203771489]
We investigate target-driven visual navigation using deep reinforcement learning (DRL) in 3D indoor scenes.
Our proposed method combines visual features and 3D spatial representations to learn navigation policy.
Our experiments, performed in the AI2-THOR, show that our model outperforms the baselines in both SR and SPL metrics.
arXiv Detail & Related papers (2020-04-29T08:46:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.