Learning to Assess Danger from Movies for Cooperative Escape Planning in
Hazardous Environments
- URL: http://arxiv.org/abs/2207.13791v1
- Date: Wed, 27 Jul 2022 21:07:15 GMT
- Title: Learning to Assess Danger from Movies for Cooperative Escape Planning in
Hazardous Environments
- Authors: Vikram Shree, Sarah Allen, Beatriz Asfora, Jacopo Banfi, Mark Campbell
- Abstract summary: It is difficult to replicate such scenarios in the real world, which is necessary for training and testing purposes.
Current systems are not fully able to take advantage of the rich multi-modal data available in such hazardous environments.
We propose to harness the enormous amount of visual content available in the form of movies and TV shows, and develop a dataset that can represent hazardous environments encountered in the real world.
- Score: 4.042350304426974
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: There has been a plethora of work towards improving robot perception and
navigation, yet their application in hazardous environments, like during a fire
or an earthquake, is still at a nascent stage. We hypothesize two key
challenges here: first, it is difficult to replicate such scenarios in the real
world, which is necessary for training and testing purposes. Second, current
systems are not fully able to take advantage of the rich multi-modal data
available in such hazardous environments. To address the first challenge, we
propose to harness the enormous amount of visual content available in the form
of movies and TV shows, and develop a dataset that can represent hazardous
environments encountered in the real world. The data is annotated with
high-level danger ratings for realistic disaster images, and corresponding
keywords are provided that summarize the content of the scene. In response to
the second challenge, we propose a multi-modal danger estimation pipeline for
collaborative human-robot escape scenarios. Our Bayesian framework improves
danger estimation by fusing information from robot's camera sensor and language
inputs from the human. Furthermore, we augment the estimation module with a
risk-aware planner that helps in identifying safer paths out of the dangerous
environment. Through extensive simulations, we exhibit the advantages of our
multi-modal perception framework that gets translated into tangible benefits
such as higher success rate in a collaborative human-robot mission.
Related papers
- Traffic and Safety Rule Compliance of Humans in Diverse Driving Situations [48.924085579865334]
Analyzing human data is crucial for developing autonomous systems that replicate safe driving practices.
This paper presents a comparative evaluation of human compliance with traffic and safety rules across multiple trajectory prediction datasets.
arXiv Detail & Related papers (2024-11-04T09:21:00Z) - ReGentS: Real-World Safety-Critical Driving Scenario Generation Made Stable [88.08120417169971]
Machine learning based autonomous driving systems often face challenges with safety-critical scenarios that are rare in real-world data.
This work explores generating safety-critical driving scenarios by modifying complex real-world regular scenarios through trajectory optimization.
Our approach addresses unrealistic diverging trajectories and unavoidable collision scenarios that are not useful for training robust planner.
arXiv Detail & Related papers (2024-09-12T08:26:33Z) - Commonsense Reasoning for Legged Robot Adaptation with Vision-Language Models [81.55156507635286]
Legged robots are physically capable of navigating a diverse variety of environments and overcoming a wide range of obstructions.
Current learning methods often struggle with generalization to the long tail of unexpected situations without heavy human supervision.
We propose a system, VLM-Predictive Control (VLM-PC), combining two key components that we find to be crucial for eliciting on-the-fly, adaptive behavior selection.
arXiv Detail & Related papers (2024-07-02T21:00:30Z) - Synthetic Multimodal Dataset for Empowering Safety and Well-being in
Home Environments [1.747623282473278]
This paper presents a synthetic multimodaltemporal of daily activities that fuses video data from a 3D virtual space simulator with knowledge graphs.
The dataset is developed for the Knowledge Graph Reasoning Challenge Social Issues (KGRC4SI), which focuses on identifying and addressing hazardous situations in the home environment.
arXiv Detail & Related papers (2024-01-26T10:05:41Z) - COPILOT: Human-Environment Collision Prediction and Localization from
Egocentric Videos [62.34712951567793]
The ability to forecast human-environment collisions from egocentric observations is vital to enable collision avoidance in applications such as VR, AR, and wearable assistive robotics.
We introduce the challenging problem of predicting collisions in diverse environments from multi-view egocentric videos captured from body-mounted cameras.
We propose a transformer-based model called COPILOT to perform collision prediction and localization simultaneously.
arXiv Detail & Related papers (2022-10-04T17:49:23Z) - Safe reinforcement learning of dynamic high-dimensional robotic tasks:
navigation, manipulation, interaction [31.553783147007177]
In reinforcement learning, safety is even more fundamental for exploring an environment without causing any damage.
This paper introduces a new formulation of safe exploration for reinforcement learning of various robotic tasks.
Our approach applies to a wide class of robotic platforms and enforces safety even under complex collision constraints learned from data.
arXiv Detail & Related papers (2022-09-27T11:23:49Z) - Incremental 3D Scene Completion for Safe and Efficient Exploration
Mapping and Planning [60.599223456298915]
We propose a novel way to integrate deep learning into exploration by leveraging 3D scene completion for informed, safe, and interpretable mapping and planning.
We show that our method can speed up coverage of an environment by 73% compared to the baselines with only minimal reduction in map accuracy.
Even if scene completions are not included in the final map, we show that they can be used to guide the robot to choose more informative paths, speeding up the measurement of the scene with the robot's sensors by 35%.
arXiv Detail & Related papers (2022-08-17T14:19:33Z) - Learning Risk-aware Costmaps for Traversability in Challenging
Environments [16.88528967313285]
We introduce a neural network architecture for robustly learning the distribution of traversability costs.
Because we are motivated by preserving the life of the robot, we tackle this learning problem from the perspective of learning tail-risks.
We show that this approach reliably learns the expected tail risk given a desired probability risk threshold between 0 and 1.
arXiv Detail & Related papers (2021-07-25T04:12:03Z) - Vision based Pedestrian Potential Risk Analysis based on Automated
Behavior Feature Extraction for Smart and Safe City [5.759189800028578]
We propose a comprehensive analytical model for pedestrian potential risk using video footage gathered by road security cameras deployed at such crossings.
The proposed system automatically detects vehicles and pedestrians, calculates trajectories by frames, and extracts behavioral features affecting the likelihood of potentially dangerous scenes between these objects.
We validated feasibility and applicability by applying it in multiple crosswalks in Osan city, Korea.
arXiv Detail & Related papers (2021-05-06T11:03:10Z) - Batch Exploration with Examples for Scalable Robotic Reinforcement
Learning [63.552788688544254]
Batch Exploration with Examples (BEE) explores relevant regions of the state-space guided by a modest number of human provided images of important states.
BEE is able to tackle challenging vision-based manipulation tasks both in simulation and on a real Franka robot.
arXiv Detail & Related papers (2020-10-22T17:49:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.