Combating Spatial Disorientation in a Dynamic Self-Stabilization Task Using AI Assistants
- URL: http://arxiv.org/abs/2409.14565v1
- Date: Mon, 9 Sep 2024 21:06:22 GMT
- Title: Combating Spatial Disorientation in a Dynamic Self-Stabilization Task Using AI Assistants
- Authors: Sheikh Mannan, Paige Hansen, Vivekanand Pandey Vimal, Hannah N. Davies, Paul DiZio, Nikhil Krishnaswamy,
- Abstract summary: Spatial disorientation is a leading cause of fatal aircraft accidents.
This paper explores the potential of AI agents to aid pilots in maintaining balance and preventing unrecoverable losses of control.
- Score: 5.42300240053097
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Spatial disorientation is a leading cause of fatal aircraft accidents. This paper explores the potential of AI agents to aid pilots in maintaining balance and preventing unrecoverable losses of control by offering cues and corrective measures that ameliorate spatial disorientation. A multi-axis rotation system (MARS) was used to gather data from human subjects self-balancing in a spaceflight analog condition. We trained models over this data to create "digital twins" that exemplified performance characteristics of humans with different proficiency levels. We then trained various reinforcement learning and deep learning models to offer corrective cues if loss of control is predicted. Digital twins and assistant models then co-performed a virtual inverted pendulum (VIP) programmed with identical physics. From these simulations, we picked the 5 best-performing assistants based on task metrics such as crash frequency and mean distance from the direction of balance. These were used in a co-performance study with 20 new human subjects performing a version of the VIP task with degraded spatial information. We show that certain AI assistants were able to improve human performance and that reinforcement-learning based assistants were objectively more effective but rated as less trusted and preferable by humans.
Related papers
- Mitigating Covariate Shift in Imitation Learning for Autonomous Vehicles Using Latent Space Generative World Models [60.87795376541144]
A world model is a neural network capable of predicting an agent's next state given past states and actions.
During end-to-end training, our policy learns how to recover from errors by aligning with states observed in human demonstrations.
We present qualitative and quantitative results, demonstrating significant improvements upon prior state of the art in closed-loop testing.
arXiv Detail & Related papers (2024-09-25T06:48:25Z) - Addressing and Visualizing Misalignments in Human Task-Solving Trajectories [5.166083532861163]
We propose a visualization tool and an algorithm to detect and categorize discrepancies in trajectory data.
We expect that eliminating these misalignments could significantly improve the utility of trajectory data for AI model training.
arXiv Detail & Related papers (2024-09-21T16:38:22Z) - MIRACLE: Inverse Reinforcement and Curriculum Learning Model for
Human-inspired Mobile Robot Navigation [13.824617183645291]
In emergency scenarios, mobile robots must navigate like humans, interpreting stimuli to locate potential victims rapidly without interfering with first responders.
We propose a solution, MIRACLE, that employs gamified learning to gather stimuli-driven human navigational data.
This data is then used to train a Deep Inverse Maximum Entropy Reinforcement Learning model, reducing reliance on demonstrator abilities.
arXiv Detail & Related papers (2023-12-06T18:13:21Z) - Robot Fine-Tuning Made Easy: Pre-Training Rewards and Policies for
Autonomous Real-World Reinforcement Learning [58.3994826169858]
We introduce RoboFuME, a reset-free fine-tuning system for robotic reinforcement learning.
Our insights are to utilize offline reinforcement learning techniques to ensure efficient online fine-tuning of a pre-trained policy.
Our method can incorporate data from an existing robot dataset and improve on a target task within as little as 3 hours of autonomous real-world experience.
arXiv Detail & Related papers (2023-10-23T17:50:08Z) - Multi-model fusion for Aerial Vision and Dialog Navigation based on
human attention aids [69.98258892165767]
We present an aerial navigation task for the 2023 ICCV Conversation History.
We propose an effective method of fusion training of Human Attention Aided Transformer model (HAA-Transformer) and Human Attention Aided LSTM (HAA-LSTM) models.
arXiv Detail & Related papers (2023-08-27T10:32:52Z) - Multi-Agent Deep Reinforcement Learning for Dynamic Avatar Migration in
AIoT-enabled Vehicular Metaverses with Trajectory Prediction [70.9337170201739]
We propose a model to predict the future trajectories of intelligent vehicles based on their historical data.
We show that our proposed algorithm can effectively reduce the latency of executing avatar tasks by around 25% without prediction.
arXiv Detail & Related papers (2023-06-26T13:27:11Z) - Improving GANs with A Dynamic Discriminator [106.54552336711997]
We argue that a discriminator with an on-the-fly adjustment on its capacity can better accommodate such a time-varying task.
A comprehensive empirical study confirms that the proposed training strategy, termed as DynamicD, improves the synthesis performance without incurring any additional cost or training objectives.
arXiv Detail & Related papers (2022-09-20T17:57:33Z) - OSCAR: Data-Driven Operational Space Control for Adaptive and Robust
Robot Manipulation [50.59541802645156]
Operational Space Control (OSC) has been used as an effective task-space controller for manipulation.
We propose OSC for Adaptation and Robustness (OSCAR), a data-driven variant of OSC that compensates for modeling errors.
We evaluate our method on a variety of simulated manipulation problems, and find substantial improvements over an array of controller baselines.
arXiv Detail & Related papers (2021-10-02T01:21:38Z) - Maneuver Identification Challenge [6.154050700252063]
The Maneuver Identification Challenge provides thousands of trajectories collected from pilots practicing in flight simulators.
The first challenge is separating physically plausible (good) trajectories from unfeasible (bad) trajectories.
The second challenge is to label trajectories with their intended maneuvers and to assess the quality of those maneuvers.
arXiv Detail & Related papers (2021-08-25T22:41:45Z) - Adaptive t-Momentum-based Optimization for Unknown Ratio of Outliers in
Amateur Data in Imitation Learning [3.145455301228175]
Behavioral (BC) bears a high potential for safe and direct transfer of human skills to robots.
In order to allow the imitators to effectively learn from imperfect demonstrations, we propose to employ the robust t-momentum optimization algorithm.
We show empirically how the algorithm can be used to produce robust BC imitators against datasets with unknown heaviness.
arXiv Detail & Related papers (2021-08-02T04:30:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.