Deep reinforcement learning with a particle dynamics environment applied
to emergency evacuation of a room with obstacles
- URL: http://arxiv.org/abs/2012.00065v1
- Date: Mon, 30 Nov 2020 19:34:57 GMT
- Title: Deep reinforcement learning with a particle dynamics environment applied
to emergency evacuation of a room with obstacles
- Authors: Yihao Zhang, Zhaojie Chai and George Lykotrafitis
- Abstract summary: We develop a deep reinforcement learning algorithm in association with the social force model to train agents to find the fastest evacuation path.
We first show that in the case of a room without obstacles the resulting self-driven force points directly towards the exit as in the social force model.
We show that our method produces similar results with the social force model when the obstacle is convex.
- Score: 3.031582944011582
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: A very successful model for simulating emergency evacuation is the
social-force model. At the heart of the model is the self-driven force that is
applied to an agent and is directed towards the exit. However, it is not clear
if the application of this force results in optimal evacuation, especially in
complex environments with obstacles. Here, we develop a deep reinforcement
learning algorithm in association with the social force model to train agents
to find the fastest evacuation path. During training, we penalize every step of
an agent in the room and give zero reward at the exit. We adopt the Dyna-Q
learning approach. We first show that in the case of a room without obstacles
the resulting self-driven force points directly towards the exit as in the
social force model and that the median exit time intervals calculated using the
two methods are not significantly different. Then, we investigate evacuation of
a room with one obstacle and one exit. We show that our method produces similar
results with the social force model when the obstacle is convex. However, in
the case of concave obstacles, which sometimes can act as traps for agents
governed purely by the social force model and prohibit complete room
evacuation, our approach is clearly advantageous since it derives a policy that
results in object avoidance and complete room evacuation without additional
assumptions. We also study evacuation of a room with multiple exits. We show
that agents are able to evacuate efficiently from the nearest exit through a
shared network trained for a single agent. Finally, we test the robustness of
the Dyna-Q learning approach in a complex environment with multiple exits and
obstacles. Overall, we show that our model can efficiently simulate emergency
evacuation in complex environments with multiple room exits and obstacles where
it is difficult to obtain an intuitive rule for fast evacuation.
Related papers
- Enhancing Building Safety Design for Active Shooter Incidents: Exploration of Building Exit Parameters using Reinforcement Learning-Based Simulations [1.3374504717801061]
This study proposes a reinforcement learning-based simulation approach addressing gaps in existing research.
We developed an autonomous agent to simulate an active shooter within a realistic office environment.
arXiv Detail & Related papers (2024-07-15T05:08:38Z) - RACER: Epistemic Risk-Sensitive RL Enables Fast Driving with Fewer Crashes [57.319845580050924]
We propose a reinforcement learning framework that combines risk-sensitive control with an adaptive action space curriculum.
We show that our algorithm is capable of learning high-speed policies for a real-world off-road driving task.
arXiv Detail & Related papers (2024-05-07T23:32:36Z) - Denoising Heat-inspired Diffusion with Insulators for Collision Free
Motion Planning [3.074694788117593]
Diffusion models have risen as a powerful tool in robotics due to their flexibility and multi-modality.
We present a method that simultaneously generates only reachable goals and plans motions that avoid obstacles.
arXiv Detail & Related papers (2023-10-19T09:39:07Z) - AI planning in the imagination: High-level planning on learned abstract
search spaces [68.75684174531962]
We propose a new method, called PiZero, that gives an agent the ability to plan in an abstract search space that the agent learns during training.
We evaluate our method on multiple domains, including the traveling salesman problem, Sokoban, 2048, the facility location problem, and Pacman.
arXiv Detail & Related papers (2023-08-16T22:47:16Z) - Enhanced method for reinforcement learning based dynamic obstacle
avoidance by assessment of collision risk [0.0]
This paper proposes a general training environment where we gain control over the difficulty of the obstacle avoidance task.
We found that shifting the training towards a greater task difficulty can massively increase the final performance.
arXiv Detail & Related papers (2022-12-08T07:46:42Z) - Optimizing Trajectories for Highway Driving with Offline Reinforcement
Learning [11.970409518725491]
We propose a Reinforcement Learning-based approach to autonomous driving.
We compare the performance of our agent against four other highway driving agents.
We demonstrate that our offline trained agent, with randomly collected data, learns to drive smoothly, achieving as close as possible to the desired velocity, while outperforming the other agents.
arXiv Detail & Related papers (2022-03-21T13:13:08Z) - Online reinforcement learning with sparse rewards through an active
inference capsule [62.997667081978825]
This paper introduces an active inference agent which minimizes the novel free energy of the expected future.
Our model is capable of solving sparse-reward problems with a very high sample efficiency.
We also introduce a novel method for approximating the prior model from the reward function, which simplifies the expression of complex objectives.
arXiv Detail & Related papers (2021-06-04T10:03:36Z) - Vision-Based Mobile Robotics Obstacle Avoidance With Deep Reinforcement
Learning [49.04274612323564]
Obstacle avoidance is a fundamental and challenging problem for autonomous navigation of mobile robots.
In this paper, we consider the problem of obstacle avoidance in simple 3D environments where the robot has to solely rely on a single monocular camera.
We tackle the obstacle avoidance problem as a data-driven end-to-end deep learning approach.
arXiv Detail & Related papers (2021-03-08T13:05:46Z) - Learning-'N-Flying: A Learning-based, Decentralized Mission Aware UAS
Collision Avoidance Scheme [0.0]
Learning-'N-Flying (LNF) is a multi-UAS Collision Avoidance (CA) framework.
It is decentralized, works on-the-fly and allows autonomous UAS managed by different operators to safely carry out complex missions.
We show that our method can run online (computation time in the order of milliseconds), and under certain assumptions has failure rates of less than 1% in the worst-case.
arXiv Detail & Related papers (2021-01-25T20:38:17Z) - Congestion-aware Evacuation Routing using Augmented Reality Devices [96.68280427555808]
We present a congestion-aware routing solution for indoor evacuation, which produces real-time individual-customized evacuation routes among multiple destinations.
A population density map, obtained on-the-fly by aggregating locations of evacuees from user-end Augmented Reality (AR) devices, is used to model the congestion distribution inside a building.
arXiv Detail & Related papers (2020-04-25T22:54:35Z) - State-Only Imitation Learning for Dexterous Manipulation [63.03621861920732]
In this paper, we explore state-only imitation learning.
We train an inverse dynamics model and use it to predict actions for state-only demonstrations.
Our method performs on par with state-action approaches and considerably outperforms RL alone.
arXiv Detail & Related papers (2020-04-07T17:57:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.