Synthetic Multimodal Dataset for Empowering Safety and Well-being in
Home Environments
- URL: http://arxiv.org/abs/2401.14743v1
- Date: Fri, 26 Jan 2024 10:05:41 GMT
- Title: Synthetic Multimodal Dataset for Empowering Safety and Well-being in
Home Environments
- Authors: Takanori Ugai, Shusaku Egami, Swe Nwe Nwe Htun, Kouji Kozaki, Takahiro
Kawamura, Ken Fukuda
- Abstract summary: This paper presents a synthetic multimodaltemporal of daily activities that fuses video data from a 3D virtual space simulator with knowledge graphs.
The dataset is developed for the Knowledge Graph Reasoning Challenge Social Issues (KGRC4SI), which focuses on identifying and addressing hazardous situations in the home environment.
- Score: 1.747623282473278
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This paper presents a synthetic multimodal dataset of daily activities that
fuses video data from a 3D virtual space simulator with knowledge graphs
depicting the spatiotemporal context of the activities. The dataset is
developed for the Knowledge Graph Reasoning Challenge for Social Issues
(KGRC4SI), which focuses on identifying and addressing hazardous situations in
the home environment. The dataset is available to the public as a valuable
resource for researchers and practitioners developing innovative solutions
recognizing human behaviors to enhance safety and well-being in
Related papers
- Can-Do! A Dataset and Neuro-Symbolic Grounded Framework for Embodied Planning with Large Multimodal Models [85.55649666025926]
We introduce Can-Do, a benchmark dataset designed to evaluate embodied planning abilities.
Our dataset includes 400 multimodal samples, each consisting of natural language user instructions, visual images depicting the environment, state changes, and corresponding action plans.
We propose NeuroGround, a neurosymbolic framework that first grounds the plan generation in the perceived environment states and then leverages symbolic planning engines to augment the model-generated plans.
arXiv Detail & Related papers (2024-09-22T00:30:11Z) - Enhancing Generalizability of Representation Learning for Data-Efficient 3D Scene Understanding [50.448520056844885]
We propose a generative Bayesian network to produce diverse synthetic scenes with real-world patterns.
A series of experiments robustly display our method's consistent superiority over existing state-of-the-art pre-training approaches.
arXiv Detail & Related papers (2024-06-17T07:43:53Z) - Best Practices and Lessons Learned on Synthetic Data [83.63271573197026]
The success of AI models relies on the availability of large, diverse, and high-quality datasets.
Synthetic data has emerged as a promising solution by generating artificial data that mimics real-world patterns.
arXiv Detail & Related papers (2024-04-11T06:34:17Z) - Object Detectors in the Open Environment: Challenges, Solutions, and Outlook [95.3317059617271]
The dynamic and intricate nature of the open environment poses novel and formidable challenges to object detectors.
This paper aims to conduct a comprehensive review and analysis of object detectors in open environments.
We propose a framework that includes four quadrants (i.e., out-of-domain, out-of-category, robust learning, and incremental learning) based on the dimensions of the data / target changes.
arXiv Detail & Related papers (2024-03-24T19:32:39Z) - LiveHPS: LiDAR-based Scene-level Human Pose and Shape Estimation in Free
Environment [59.320414108383055]
We present LiveHPS, a novel single-LiDAR-based approach for scene-level human pose and shape estimation.
We propose a huge human motion dataset, named FreeMotion, which is collected in various scenarios with diverse human poses.
arXiv Detail & Related papers (2024-02-27T03:08:44Z) - Big Data and Deep Learning in Smart Cities: A Comprehensive Dataset for
AI-Driven Traffic Accident Detection and Computer Vision Systems [0.0]
This study delves into the application of cutting-edge technological methods in smart cities.
We present a novel comprehensive dataset for traffic accident detection.
This dataset is expected to advance academic research and also enhance real-time accident detection applications.
arXiv Detail & Related papers (2024-01-07T21:50:24Z) - Synthesizing Event-centric Knowledge Graphs of Daily Activities Using
Virtual Space [0.3324876873771104]
This study proposes the VirtualHome2KG framework to generate synthetic KGs of daily life activities in virtual space.
This framework augments both the synthetic video data of daily activities and the contextual semantic data corresponding to the video contents.
Various applications that have conventionally been difficult to develop due to the insufficient availability of relevant data and semantic information can be developed.
arXiv Detail & Related papers (2023-07-30T11:50:36Z) - Towards Ubiquitous Semantic Metaverse: Challenges, Approaches, and
Opportunities [68.03971716740823]
In recent years, ubiquitous semantic Metaverse has been studied to revolutionize immersive cyber-virtual experiences for augmented reality (AR) and virtual reality (VR) users.
This survey focuses on the representation and intelligence for the four fundamental system components in ubiquitous Metaverse.
arXiv Detail & Related papers (2023-07-13T11:14:46Z) - Socially Enhanced Situation Awareness from Microblogs using Artificial
Intelligence: A Survey [0.2320417845168326]
The rise of social media platforms provides an infinitely rich source of aggregate knowledge of the world around us.
We focus on microblog social media data with applications to Situation Awareness.
We provide a novel, unified methodological perspective, identify key results and challenges.
arXiv Detail & Related papers (2022-09-13T04:03:19Z) - Learning to Assess Danger from Movies for Cooperative Escape Planning in
Hazardous Environments [4.042350304426974]
It is difficult to replicate such scenarios in the real world, which is necessary for training and testing purposes.
Current systems are not fully able to take advantage of the rich multi-modal data available in such hazardous environments.
We propose to harness the enormous amount of visual content available in the form of movies and TV shows, and develop a dataset that can represent hazardous environments encountered in the real world.
arXiv Detail & Related papers (2022-07-27T21:07:15Z) - Predicting the Future from First Person (Egocentric) Vision: A Survey [18.07516837332113]
This survey summarises the evolution of studies in the context of future prediction from egocentric vision.
It makes an overview of applications, devices, existing problems, commonly used datasets, models and input modalities.
Our analysis highlights that methods for future prediction from egocentric vision can have a significant impact in a range of applications.
arXiv Detail & Related papers (2021-07-28T14:58:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.