Let's Play for Action: Recognizing Activities of Daily Living by
Learning from Life Simulation Video Games
- URL: http://arxiv.org/abs/2107.05617v1
- Date: Mon, 12 Jul 2021 17:53:38 GMT
- Title: Let's Play for Action: Recognizing Activities of Daily Living by
Learning from Life Simulation Video Games
- Authors: Alina Roitberg, David Schneider, Aulia Djamal, Constantin Seibold,
Simon Rei{\ss}, Rainer Stiefelhagen
- Abstract summary: We introduce the SIMS4ACTION dataset created with the popular commercial game THE SIMS 4.
We build Sims4Action by specifically executing actions-of-interest in a "top-down" manner, while the gaming circumstances allow us to freely switch between environments.
We integrate two modern algorithms for video-based activity recognition in our framework, revealing the value of life simulation video games as an inexpensive and far less intrusive source of training data.
- Score: 26.79922049563356
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recognizing Activities of Daily Living (ADL) is a vital process for
intelligent assistive robots, but collecting large annotated datasets requires
time-consuming temporal labeling and raises privacy concerns, e.g., if the data
is collected in a real household. In this work, we explore the concept of
constructing training examples for ADL recognition by playing life simulation
video games and introduce the SIMS4ACTION dataset created with the popular
commercial game THE SIMS 4. We build Sims4Action by specifically executing
actions-of-interest in a "top-down" manner, while the gaming circumstances
allow us to freely switch between environments, camera angles and subject
appearances. While ADL recognition on gaming data is interesting from the
theoretical perspective, the key challenge arises from transferring it to the
real-world applications, such as smart-homes or assistive robotics. To meet
this requirement, Sims4Action is accompanied with a GamingToReal benchmark,
where the models are evaluated on real videos derived from an existing ADL
dataset. We integrate two modern algorithms for video-based activity
recognition in our framework, revealing the value of life simulation video
games as an inexpensive and far less intrusive source of training data.
However, our results also indicate that tasks involving a mixture of gaming and
real data are challenging, opening a new research direction. We will make our
dataset publicly available at https://github.com/aroitberg/sims4action.
Related papers
- Video2Policy: Scaling up Manipulation Tasks in Simulation through Internet Videos [61.925837909969815]
We introduce Video2Policy, a novel framework that leverages internet RGB videos to reconstruct tasks based on everyday human behavior.
Our method can successfully train RL policies on such tasks, including complex and challenging tasks such as throwing.
We show that the generated simulation data can be scaled up for training a general policy, and it can be transferred back to the real robot in a Real2Sim2Real way.
arXiv Detail & Related papers (2025-02-14T03:22:03Z) - Robot Learning with Super-Linear Scaling [20.730206708381704]
CASHER is a pipeline for scaling up data collection and learning in simulation where the performance scales superlinearly with human effort.
We show that CASHER enables fine-tuning of pre-trained policies to a target scenario using a video scan without any additional human effort.
arXiv Detail & Related papers (2024-12-02T18:12:02Z) - Benchmarking the Sim-to-Real Gap in Cloth Manipulation [10.530012817995656]
We present a benchmark dataset to evaluate the sim-to-real gap in cloth manipulation.
We use the dataset to evaluate the reality gap, computational time, and stability of four popular deformable object simulators.
arXiv Detail & Related papers (2023-10-14T09:36:01Z) - Learning Interactive Real-World Simulators [96.5991333400566]
We explore the possibility of learning a universal simulator of real-world interaction through generative modeling.
We use the simulator to train both high-level vision-language policies and low-level reinforcement learning policies.
Video captioning models can benefit from training with simulated experience, opening up even wider applications.
arXiv Detail & Related papers (2023-10-09T19:42:22Z) - Learning Sim-to-Real Dense Object Descriptors for Robotic Manipulation [4.7246285569677315]
We present Sim-to-Real Dense Object Nets (SRDONs), a dense object descriptor that not only understands the object via appropriate representation but also maps simulated and real data to a unified feature space with pixel consistency.
We demonstrate in experiments that pre-trained SRDONs significantly improve performances on unseen objects and unseen visual environments for various robotic tasks with zero real-world training.
arXiv Detail & Related papers (2023-04-18T02:28:55Z) - Promptable Game Models: Text-Guided Game Simulation via Masked Diffusion
Models [68.85478477006178]
We present a Promptable Game Model (PGM) for neural video game simulators.
It allows a user to play the game by prompting it with high- and low-level action sequences.
Most captivatingly, our PGM unlocks the director's mode, where the game is played by specifying goals for the agents in the form of a prompt.
Our method significantly outperforms existing neural video game simulators in terms of rendering quality and unlocks applications beyond the capabilities of the current state of the art.
arXiv Detail & Related papers (2023-03-23T17:43:17Z) - Sim2real Transfer Learning for Point Cloud Segmentation: An Industrial
Application Case on Autonomous Disassembly [55.41644538483948]
We present an industrial application case that uses sim2real transfer learning for point cloud data.
We provide insights on how to generate and process synthetic point cloud data.
A novel patch-based attention network is proposed additionally to tackle this problem.
arXiv Detail & Related papers (2023-01-12T14:00:37Z) - DeXtreme: Transfer of Agile In-hand Manipulation from Simulation to
Reality [64.51295032956118]
We train a policy that can perform robust dexterous manipulation on an anthropomorphic robot hand.
Our work reaffirms the possibilities of sim-to-real transfer for dexterous manipulation in diverse kinds of hardware and simulator setups.
arXiv Detail & Related papers (2022-10-25T01:51:36Z) - Point Cloud Based Reinforcement Learning for Sim-to-Real and Partial
Observability in Visual Navigation [62.22058066456076]
Reinforcement Learning (RL) represents powerful tools to solve complex robotic tasks.
RL does not work directly in the real-world, which is known as the sim-to-real transfer problem.
We propose a method that learns on an observation space constructed by point clouds and environment randomization.
arXiv Detail & Related papers (2020-07-27T17:46:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.