Scene-aware Generative Network for Human Motion Synthesis
- URL: http://arxiv.org/abs/2105.14804v1
- Date: Mon, 31 May 2021 09:05:50 GMT
- Title: Scene-aware Generative Network for Human Motion Synthesis
- Authors: Jingbo Wang, Sijie Yan, Bo Dai, Dahua LIn
- Abstract summary: We propose a new framework, with the interaction between the scene and the human motion taken into account.
Considering the uncertainty of human motion, we formulate this task as a generative task.
We derive a GAN based learning approach, with discriminators to enforce the compatibility between the human motion and the contextual scene.
- Score: 125.21079898942347
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We revisit human motion synthesis, a task useful in various real world
applications, in this paper. Whereas a number of methods have been developed
previously for this task, they are often limited in two aspects: focusing on
the poses while leaving the location movement behind, and ignoring the impact
of the environment on the human motion. In this paper, we propose a new
framework, with the interaction between the scene and the human motion taken
into account. Considering the uncertainty of human motion, we formulate this
task as a generative task, whose objective is to generate plausible human
motion conditioned on both the scene and the human initial position. This
framework factorizes the distribution of human motions into a distribution of
movement trajectories conditioned on scenes and that of body pose dynamics
conditioned on both scenes and trajectories. We further derive a GAN based
learning approach, with discriminators to enforce the compatibility between the
human motion and the contextual scene as well as the 3D to 2D projection
constraints. We assess the effectiveness of the proposed method on two
challenging datasets, which cover both synthetic and real world environments.
Related papers
- Multi-Condition Latent Diffusion Network for Scene-Aware Neural Human Motion Prediction [46.309401205546656]
Real-world human movements are goal-directed and highly influenced by the spatial layout of their surrounding scenes.
We propose a Multi-Condition Latent Diffusion network (MCLD) that reformulates the human motion prediction task as a multi-condition joint inference problem.
Our network achieves significant improvements over the state-of-the-art methods on both realistic and diverse predictions.
arXiv Detail & Related papers (2024-05-29T02:21:31Z) - Generating Human Motion in 3D Scenes from Text Descriptions [60.04976442328767]
This paper focuses on the task of generating human motions in 3D indoor scenes given text descriptions of the human-scene interactions.
We propose a new approach that decomposes the complex problem into two more manageable sub-problems.
For language grounding of the target object, we leverage the power of large language models; for motion generation, we design an object-centric scene representation.
arXiv Detail & Related papers (2024-05-13T14:30:12Z) - ReMoS: 3D Motion-Conditioned Reaction Synthesis for Two-Person Interactions [66.87211993793807]
We present ReMoS, a denoising diffusion based model that synthesizes full body motion of a person in two person interaction scenario.
We demonstrate ReMoS across challenging two person scenarios such as pair dancing, Ninjutsu, kickboxing, and acrobatics.
We also contribute the ReMoCap dataset for two person interactions containing full body and finger motions.
arXiv Detail & Related papers (2023-11-28T18:59:52Z) - Task-Oriented Human-Object Interactions Generation with Implicit Neural
Representations [61.659439423703155]
TOHO: Task-Oriented Human-Object Interactions Generation with Implicit Neural Representations.
Our method generates continuous motions that are parameterized only by the temporal coordinate.
This work takes a step further toward general human-scene interaction simulation.
arXiv Detail & Related papers (2023-03-23T09:31:56Z) - Towards Diverse and Natural Scene-aware 3D Human Motion Synthesis [117.15586710830489]
We focus on the problem of synthesizing diverse scene-aware human motions under the guidance of target action sequences.
Based on this factorized scheme, a hierarchical framework is proposed, with each sub-module responsible for modeling one aspect.
Experiment results show that the proposed framework remarkably outperforms previous methods in terms of diversity and naturalness.
arXiv Detail & Related papers (2022-05-25T18:20:01Z) - Action2Motion: Conditioned Generation of 3D Human Motions [28.031644518303075]
We aim to generateplausible human motion sequences in 3D.
Each sampled sequence faithfully resembles anaturalhuman bodyarticulation dynamics.
A new 3D human motion dataset, HumanAct12, is also constructed.
arXiv Detail & Related papers (2020-07-30T05:29:59Z) - Long-term Human Motion Prediction with Scene Context [60.096118270451974]
We propose a novel three-stage framework for predicting human motion.
Our method first samples multiple human motion goals, then plans 3D human paths towards each goal, and finally predicts 3D human pose sequences following each path.
arXiv Detail & Related papers (2020-07-07T17:59:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.