Let's Play for Action: Recognizing Activities of Daily Living by
Learning from Life Simulation Video Games
- URL: http://arxiv.org/abs/2107.05617v1
- Date: Mon, 12 Jul 2021 17:53:38 GMT
- Title: Let's Play for Action: Recognizing Activities of Daily Living by
Learning from Life Simulation Video Games
- Authors: Alina Roitberg, David Schneider, Aulia Djamal, Constantin Seibold,
Simon Rei{\ss}, Rainer Stiefelhagen
- Abstract summary: We introduce the SIMS4ACTION dataset created with the popular commercial game THE SIMS 4.
We build Sims4Action by specifically executing actions-of-interest in a "top-down" manner, while the gaming circumstances allow us to freely switch between environments.
We integrate two modern algorithms for video-based activity recognition in our framework, revealing the value of life simulation video games as an inexpensive and far less intrusive source of training data.
- Score: 26.79922049563356
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recognizing Activities of Daily Living (ADL) is a vital process for
intelligent assistive robots, but collecting large annotated datasets requires
time-consuming temporal labeling and raises privacy concerns, e.g., if the data
is collected in a real household. In this work, we explore the concept of
constructing training examples for ADL recognition by playing life simulation
video games and introduce the SIMS4ACTION dataset created with the popular
commercial game THE SIMS 4. We build Sims4Action by specifically executing
actions-of-interest in a "top-down" manner, while the gaming circumstances
allow us to freely switch between environments, camera angles and subject
appearances. While ADL recognition on gaming data is interesting from the
theoretical perspective, the key challenge arises from transferring it to the
real-world applications, such as smart-homes or assistive robotics. To meet
this requirement, Sims4Action is accompanied with a GamingToReal benchmark,
where the models are evaluated on real videos derived from an existing ADL
dataset. We integrate two modern algorithms for video-based activity
recognition in our framework, revealing the value of life simulation video
games as an inexpensive and far less intrusive source of training data.
However, our results also indicate that tasks involving a mixture of gaming and
real data are challenging, opening a new research direction. We will make our
dataset publicly available at https://github.com/aroitberg/sims4action.
Related papers
- RoboCasa: Large-Scale Simulation of Everyday Tasks for Generalist Robots [25.650235551519952]
We present RoboCasa, a large-scale simulation framework for training generalist robots in everyday environments.
We provide thousands of 3D assets across over 150 object categories and dozens of interactable furniture and appliances.
Our experiments show a clear scaling trend in using synthetically generated robot data for large-scale imitation learning.
arXiv Detail & Related papers (2024-06-04T17:41:31Z) - Benchmarking the Sim-to-Real Gap in Cloth Manipulation [10.530012817995656]
We present a benchmark dataset to evaluate the sim-to-real gap in cloth manipulation.
We use the dataset to evaluate the reality gap, computational time, and stability of four popular deformable object simulators.
arXiv Detail & Related papers (2023-10-14T09:36:01Z) - Learning Interactive Real-World Simulators [96.5991333400566]
We explore the possibility of learning a universal simulator of real-world interaction through generative modeling.
We use the simulator to train both high-level vision-language policies and low-level reinforcement learning policies.
Video captioning models can benefit from training with simulated experience, opening up even wider applications.
arXiv Detail & Related papers (2023-10-09T19:42:22Z) - Learning Sim-to-Real Dense Object Descriptors for Robotic Manipulation [4.7246285569677315]
We present Sim-to-Real Dense Object Nets (SRDONs), a dense object descriptor that not only understands the object via appropriate representation but also maps simulated and real data to a unified feature space with pixel consistency.
We demonstrate in experiments that pre-trained SRDONs significantly improve performances on unseen objects and unseen visual environments for various robotic tasks with zero real-world training.
arXiv Detail & Related papers (2023-04-18T02:28:55Z) - Promptable Game Models: Text-Guided Game Simulation via Masked Diffusion
Models [68.85478477006178]
We present a Promptable Game Model (PGM) for neural video game simulators.
It allows a user to play the game by prompting it with high- and low-level action sequences.
Most captivatingly, our PGM unlocks the director's mode, where the game is played by specifying goals for the agents in the form of a prompt.
Our method significantly outperforms existing neural video game simulators in terms of rendering quality and unlocks applications beyond the capabilities of the current state of the art.
arXiv Detail & Related papers (2023-03-23T17:43:17Z) - Sim2real Transfer Learning for Point Cloud Segmentation: An Industrial
Application Case on Autonomous Disassembly [55.41644538483948]
We present an industrial application case that uses sim2real transfer learning for point cloud data.
We provide insights on how to generate and process synthetic point cloud data.
A novel patch-based attention network is proposed additionally to tackle this problem.
arXiv Detail & Related papers (2023-01-12T14:00:37Z) - DeXtreme: Transfer of Agile In-hand Manipulation from Simulation to
Reality [64.51295032956118]
We train a policy that can perform robust dexterous manipulation on an anthropomorphic robot hand.
Our work reaffirms the possibilities of sim-to-real transfer for dexterous manipulation in diverse kinds of hardware and simulator setups.
arXiv Detail & Related papers (2022-10-25T01:51:36Z) - Point Cloud Based Reinforcement Learning for Sim-to-Real and Partial
Observability in Visual Navigation [62.22058066456076]
Reinforcement Learning (RL) represents powerful tools to solve complex robotic tasks.
RL does not work directly in the real-world, which is known as the sim-to-real transfer problem.
We propose a method that learns on an observation space constructed by point clouds and environment randomization.
arXiv Detail & Related papers (2020-07-27T17:46:59Z) - Learning to Simulate Dynamic Environments with GameGAN [109.25308647431952]
In this paper, we aim to learn a simulator by simply watching an agent interact with an environment.
We introduce GameGAN, a generative model that learns to visually imitate a desired game by ingesting screenplay and keyboard actions during training.
arXiv Detail & Related papers (2020-05-25T14:10:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.