Learning Disentangled Representations for Controllable Human Motion
Prediction
- URL: http://arxiv.org/abs/2207.01388v1
- Date: Mon, 4 Jul 2022 13:11:11 GMT
- Title: Learning Disentangled Representations for Controllable Human Motion
Prediction
- Authors: Chunzhi Gu, Jun Yu and Chao Zhang
- Abstract summary: We propose a novel framework to learn disentangled representations for controllable human motion prediction.
Our approach is capable of predicting state-of-the-art controllable human motions both qualitatively and quantitatively.
- Score: 24.004809791890445
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Generative model-based motion prediction techniques have recently realized
predicting controlled human motions, such as predicting multiple upper human
body motions with similar lower-body motions. However, to achieve this, the
state-of-the-art methods require either subsequently learning mapping functions
to seek similar motions or training the model repetitively to enable control
over the desired portion of body. In this paper, we propose a novel framework
to learn disentangled representations for controllable human motion prediction.
Our network involves a conditional variational auto-encoder (CVAE) architecture
to model full-body human motion, and an extra CVAE path to learn only the
corresponding partial-body (e.g., lower-body) motion. Specifically, the
inductive bias imposed by the extra CVAE path encourages two latent variables
in two paths to respectively govern separate representations for each
partial-body motion. With a single training, our model is able to provide two
types of controls for the generated human motions: (i) strictly controlling one
portion of human body and (ii) adaptively controlling the other portion, by
sampling from a pair of latent spaces. Additionally, we extend and adapt a
sampling strategy to our trained model to diversify the controllable
predictions. Our framework also potentially allows new forms of control by
flexibly customizing the input for the extra CVAE path. Extensive experimental
results and ablation studies demonstrate that our approach is capable of
predicting state-of-the-art controllable human motions both qualitatively and
quantitatively.
Related papers
- CAR: Controllable Autoregressive Modeling for Visual Generation [100.33455832783416]
Controllable AutoRegressive Modeling (CAR) is a novel, plug-and-play framework that integrates conditional control into multi-scale latent variable modeling.
CAR progressively refines and captures control representations, which are injected into each autoregressive step of the pre-trained model to guide the generation process.
Our approach demonstrates excellent controllability across various types of conditions and delivers higher image quality compared to previous methods.
arXiv Detail & Related papers (2024-10-07T00:55:42Z) - InterControl: Zero-shot Human Interaction Generation by Controlling Every Joint [67.6297384588837]
We introduce a novel controllable motion generation method, InterControl, to encourage the synthesized motions maintaining the desired distance between joint pairs.
We demonstrate that the distance between joint pairs for human-wise interactions can be generated using an off-the-shelf Large Language Model.
arXiv Detail & Related papers (2023-11-27T14:32:33Z) - Executing your Commands via Motion Diffusion in Latent Space [51.64652463205012]
We propose a Motion Latent-based Diffusion model (MLD) to produce vivid motion sequences conforming to the given conditional inputs.
Our MLD achieves significant improvements over the state-of-the-art methods among extensive human motion generation tasks.
arXiv Detail & Related papers (2022-12-08T03:07:00Z) - DMMGAN: Diverse Multi Motion Prediction of 3D Human Joints using
Attention-Based Generative Adverserial Network [9.247294820004143]
We propose a transformer-based generative model for forecasting multiple diverse human motions.
Our model first predicts the pose of the body relative to the hip joint. Then the textitHip Prediction Module predicts the trajectory of the hip movement for each predicted pose frame.
We show that our system outperforms the state-of-the-art in human motion prediction while it can predict diverse multi-motion future trajectories with hip movements.
arXiv Detail & Related papers (2022-09-13T23:22:33Z) - Weakly-supervised Action Transition Learning for Stochastic Human Motion
Prediction [81.94175022575966]
We introduce the task of action-driven human motion prediction.
It aims to predict multiple plausible future motions given a sequence of action labels and a short motion history.
arXiv Detail & Related papers (2022-05-31T08:38:07Z) - Generating Smooth Pose Sequences for Diverse Human Motion Prediction [90.45823619796674]
We introduce a unified deep generative network for both diverse and controllable motion prediction.
Our experiments on two standard benchmark datasets, Human3.6M and HumanEva-I, demonstrate that our approach outperforms the state-of-the-art baselines in terms of both sample diversity and accuracy.
arXiv Detail & Related papers (2021-08-19T00:58:00Z) - TRiPOD: Human Trajectory and Pose Dynamics Forecasting in the Wild [77.59069361196404]
TRiPOD is a novel method for predicting body dynamics based on graph attentional networks.
To incorporate a real-world challenge, we learn an indicator representing whether an estimated body joint is visible/invisible at each frame.
Our evaluation shows that TRiPOD outperforms all prior work and state-of-the-art specifically designed for each of the trajectory and pose forecasting tasks.
arXiv Detail & Related papers (2021-04-08T20:01:00Z) - Learning Multiscale Correlations for Human Motion Prediction [10.335804615372629]
We propose a novel multiscale graph convolution network (MGCN) to capture the correlations among human body components.
We evaluate our approach on two standard benchmark datasets for human motion prediction.
arXiv Detail & Related papers (2021-03-19T07:58:16Z) - Aggregated Multi-GANs for Controlled 3D Human Motion Prediction [39.84142236414808]
We propose a generalization of the human motion prediction task in which control parameters can be readily incorporated to adjust the forecasted motion.
Our method is compelling in that it enables manipulable motion prediction across activity types and allows customization of the human movement in a variety of fine-grained ways.
arXiv Detail & Related papers (2021-03-17T16:22:36Z) - Reinforcement Learning Control of a Biomechanical Model of the Upper
Extremity [0.0]
We learn a control policy using a motor babbling approach as implemented in reinforcement learning.
We use a state-of-the-art biomechanical model, which includes seven actuated degrees of freedom.
To deal with the curse of dimensionality, we use a simplified second-order muscle model, acting at each degree of freedom instead of individual muscles.
arXiv Detail & Related papers (2020-11-13T19:49:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.