Diverse Human Motion Prediction Guided by Multi-Level Spatial-Temporal
Anchors
- URL: http://arxiv.org/abs/2302.04860v1
- Date: Thu, 9 Feb 2023 18:58:07 GMT
- Title: Diverse Human Motion Prediction Guided by Multi-Level Spatial-Temporal
Anchors
- Authors: Sirui Xu, Yu-Xiong Wang, Liang-Yan Gui
- Abstract summary: We propose a simple yet effective approach that disentangles randomly sampled codes with a deterministic learnable component named anchors to promote sample precision and diversity.
In principle, our spatial-temporal anchor-based sampling (STARS) can be applied to different motion predictors.
- Score: 21.915057426589744
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Predicting diverse human motions given a sequence of historical poses has
received increasing attention. Despite rapid progress, existing work captures
the multi-modal nature of human motions primarily through likelihood-based
sampling, where the mode collapse has been widely observed. In this paper, we
propose a simple yet effective approach that disentangles randomly sampled
codes with a deterministic learnable component named anchors to promote sample
precision and diversity. Anchors are further factorized into spatial anchors
and temporal anchors, which provide attractively interpretable control over
spatial-temporal disparity. In principle, our spatial-temporal anchor-based
sampling (STARS) can be applied to different motion predictors. Here we propose
an interaction-enhanced spatial-temporal graph convolutional network (IE-STGCN)
that encodes prior knowledge of human motions (e.g., spatial locality), and
incorporate the anchors into it. Extensive experiments demonstrate that our
approach outperforms state of the art in both stochastic and deterministic
prediction, suggesting it as a unified framework for modeling human motions.
Our code and pretrained models are available at
https://github.com/Sirui-Xu/STARS.
Related papers
- Multiscale Residual Learning of Graph Convolutional Sequence Chunks for
Human Motion Prediction [23.212848643552395]
A new method is proposed for human motion prediction by learning temporal and spatial dependencies.
Our proposed method is able to effectively model the sequence information for motion prediction and outperform other techniques to set a new state-of-the-art.
arXiv Detail & Related papers (2023-08-31T15:23:33Z) - Spatio-Temporal Branching for Motion Prediction using Motion Increments [55.68088298632865]
Human motion prediction (HMP) has emerged as a popular research topic due to its diverse applications.
Traditional methods rely on hand-crafted features and machine learning techniques.
We propose a noveltemporal-temporal branching network using incremental information for HMP.
arXiv Detail & Related papers (2023-08-02T12:04:28Z) - A Neuro-Symbolic Approach for Enhanced Human Motion Prediction [5.742409080817885]
We propose a neuro-symbolic approach for human motion prediction (NeuroSyM)
NeuroSyM weights differently the interactions in the neighbourhood by leveraging an intuitive technique for spatial representation called qualitative Trajectory Calculus (QTC)
Experimental results show that the NeuroSyM approach outperforms in most cases the baseline architectures in terms of prediction accuracy.
arXiv Detail & Related papers (2023-04-23T20:11:40Z) - Stochastic Trajectory Prediction via Motion Indeterminacy Diffusion [88.45326906116165]
We present a new framework to formulate the trajectory prediction task as a reverse process of motion indeterminacy diffusion (MID)
We encode the history behavior information and the social interactions as a state embedding and devise a Transformer-based diffusion model to capture the temporal dependencies of trajectories.
Experiments on the human trajectory prediction benchmarks including the Stanford Drone and ETH/UCY datasets demonstrate the superiority of our method.
arXiv Detail & Related papers (2022-03-25T16:59:08Z) - Skeleton-Based Action Segmentation with Multi-Stage Spatial-Temporal
Graph Convolutional Neural Networks [0.5156484100374059]
State-of-the-art action segmentation approaches use multiple stages of temporal convolutions.
We present multi-stage spatial-temporal graph convolutional neural networks (MS-GCN)
We replace the initial stage of temporal convolutions with spatial-temporal graph convolutions, which better exploit the spatial configuration of the joints.
arXiv Detail & Related papers (2022-02-03T17:42:04Z) - Motion Prediction via Joint Dependency Modeling in Phase Space [40.54430409142653]
We introduce a novel convolutional neural model to leverage explicit prior knowledge of motion anatomy.
We then propose a global optimization module that learns the implicit relationships between individual joint features.
Our method is evaluated on large-scale 3D human motion benchmark datasets.
arXiv Detail & Related papers (2022-01-07T08:30:01Z) - Investigating Pose Representations and Motion Contexts Modeling for 3D
Motion Prediction [63.62263239934777]
We conduct an indepth study on various pose representations with a focus on their effects on the motion prediction task.
We propose a novel RNN architecture termed AHMR (Attentive Hierarchical Motion Recurrent network) for motion prediction.
Our approach outperforms the state-of-the-art methods in short-term prediction and achieves much enhanced long-term prediction proficiency.
arXiv Detail & Related papers (2021-12-30T10:45:22Z) - Generating Smooth Pose Sequences for Diverse Human Motion Prediction [90.45823619796674]
We introduce a unified deep generative network for both diverse and controllable motion prediction.
Our experiments on two standard benchmark datasets, Human3.6M and HumanEva-I, demonstrate that our approach outperforms the state-of-the-art baselines in terms of both sample diversity and accuracy.
arXiv Detail & Related papers (2021-08-19T00:58:00Z) - Long Term Motion Prediction Using Keyposes [122.22758311506588]
We argue that, to achieve long term forecasting, predicting human pose at every time instant is unnecessary.
We call such poses "keyposes", and approximate complex motions by linearly interpolating between subsequent keyposes.
We show that learning the sequence of such keyposes allows us to predict very long term motion, up to 5 seconds in the future.
arXiv Detail & Related papers (2020-12-08T20:45:51Z) - Motion Prediction Using Temporal Inception Module [96.76721173517895]
We propose a Temporal Inception Module (TIM) to encode human motion.
Our framework produces input embeddings using convolutional layers, by using different kernel sizes for different input lengths.
The experimental results on standard motion prediction benchmark datasets Human3.6M and CMU motion capture dataset show that our approach consistently outperforms the state of the art methods.
arXiv Detail & Related papers (2020-10-06T20:26:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.