MultiAct: Long-Term 3D Human Motion Generation from Multiple Action
Labels
- URL: http://arxiv.org/abs/2212.05897v1
- Date: Mon, 12 Dec 2022 13:52:53 GMT
- Title: MultiAct: Long-Term 3D Human Motion Generation from Multiple Action
Labels
- Authors: Taeryung Lee, Gyeongsik Moon, and Kyoung Mu Lee
- Abstract summary: We present MultiAct, the first framework to generate long-term 3D human motion from multiple action labels.
It takes account of both action and motion conditions with a unified recurrent generation system.
As a result, MultiAct produces realistic long-term motion controlled by the given sequence of multiple action labels.
- Score: 59.53048564128578
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: We tackle the problem of generating long-term 3D human motion from multiple
action labels. Two main previous approaches, such as action- and
motion-conditioned methods, have limitations to solve this problem. The
action-conditioned methods generate a sequence of motion from a single action.
Hence, it cannot generate long-term motions composed of multiple actions and
transitions between actions. Meanwhile, the motion-conditioned methods generate
future motions from initial motion. The generated future motions only depend on
the past, so they are not controllable by the user's desired actions. We
present MultiAct, the first framework to generate long-term 3D human motion
from multiple action labels. MultiAct takes account of both action and motion
conditions with a unified recurrent generation system. It repetitively takes
the previous motion and action label; then, it generates a smooth transition
and the motion of the given action. As a result, MultiAct produces realistic
long-term motion controlled by the given sequence of multiple action labels.
The code will be released.
Related papers
- InfiniMotion: Mamba Boosts Memory in Transformer for Arbitrary Long Motion Generation [31.775481455602634]
Current methods struggle to handle long motion sequences as a single input due to high computational cost.
We propose InfiniMotion, a method that generates continuous motion sequences of arbitrary length within an autoregressive framework.
We highlight its groundbreaking capability by generating a continuous 1-hour human motion with around 80,000 frames.
arXiv Detail & Related papers (2024-07-14T03:12:19Z) - CoMo: Controllable Motion Generation through Language Guided Pose Code Editing [57.882299081820626]
We introduce CoMo, a Controllable Motion generation model, adept at accurately generating and editing motions.
CoMo decomposes motions into discrete and semantically meaningful pose codes.
It autoregressively generates sequences of pose codes, which are then decoded into 3D motions.
arXiv Detail & Related papers (2024-03-20T18:11:10Z) - Synthesizing Long-Term Human Motions with Diffusion Models via Coherent
Sampling [74.62570964142063]
Text-to-motion generation has gained increasing attention, but most existing methods are limited to generating short-term motions.
We propose a novel approach that utilizes a past-conditioned diffusion model with two optional coherent sampling methods.
Our proposed method is capable of generating compositional and coherent long-term 3D human motions controlled by a user-instructed long text stream.
arXiv Detail & Related papers (2023-08-03T16:18:32Z) - NEURAL MARIONETTE: A Transformer-based Multi-action Human Motion
Synthesis System [51.43113919042621]
We present a neural network-based system for long-term, multi-action human motion synthesis.
The system can produce meaningful motions with smooth transitions from simple user input.
We also present a new dataset dedicated to the multi-action motion synthesis task.
arXiv Detail & Related papers (2022-09-27T07:10:20Z) - TEACH: Temporal Action Composition for 3D Humans [50.97135662063117]
Given a series of natural language descriptions, our task is to generate 3D human motions that correspond semantically to the text.
In particular, our goal is to enable the synthesis of a series of actions, which we refer to as temporal action composition.
arXiv Detail & Related papers (2022-09-09T00:33:40Z) - MotionDiffuse: Text-Driven Human Motion Generation with Diffusion Model [35.32967411186489]
MotionDiffuse is a diffusion model-based text-driven motion generation framework.
It excels at modeling complicated data distribution and generating vivid motion sequences.
It responds to fine-grained instructions on body parts, and arbitrary-length motion synthesis with time-varied text prompts.
arXiv Detail & Related papers (2022-08-31T17:58:54Z) - Weakly-supervised Action Transition Learning for Stochastic Human Motion
Prediction [81.94175022575966]
We introduce the task of action-driven human motion prediction.
It aims to predict multiple plausible future motions given a sequence of action labels and a short motion history.
arXiv Detail & Related papers (2022-05-31T08:38:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.