Adversarial Attention for Human Motion Synthesis
- URL: http://arxiv.org/abs/2204.11751v1
- Date: Mon, 25 Apr 2022 16:12:42 GMT
- Title: Adversarial Attention for Human Motion Synthesis
- Authors: Matthew Malek-Podjaski, Fani Deligianni
- Abstract summary: We present a novel method for controllable human motion synthesis by applying attention-based probabilistic deep adversarial models with end-to-end training.
We show that we can generate synthetic human motion over both short- and long-time horizons through the use of adversarial attention.
- Score: 3.9378507882929563
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Analysing human motions is a core topic of interest for many disciplines,
from Human-Computer Interaction, to entertainment, Virtual Reality and
healthcare. Deep learning has achieved impressive results in capturing human
pose in real-time. On the other hand, due to high inter-subject variability,
human motion analysis models often suffer from not being able to generalise to
data from unseen subjects due to very limited specialised datasets available in
fields such as healthcare. However, acquiring human motion datasets is highly
time-consuming, challenging, and expensive. Hence, human motion synthesis is a
crucial research problem within deep learning and computer vision. We present a
novel method for controllable human motion synthesis by applying
attention-based probabilistic deep adversarial models with end-to-end training.
We show that we can generate synthetic human motion over both short- and
long-time horizons through the use of adversarial attention. Furthermore, we
show that we can improve the classification performance of deep learning models
in cases where there is inadequate real data, by supplementing existing
datasets with synthetic motions.
Related papers
- Scaling Up Dynamic Human-Scene Interaction Modeling [58.032368564071895]
TRUMANS is the most comprehensive motion-captured HSI dataset currently available.
It intricately captures whole-body human motions and part-level object dynamics.
We devise a diffusion-based autoregressive model that efficiently generates HSI sequences of any length.
arXiv Detail & Related papers (2024-03-13T15:45:04Z) - InterControl: Zero-shot Human Interaction Generation by Controlling Every Joint [67.6297384588837]
We introduce a novel controllable motion generation method, InterControl, to encourage the synthesized motions maintaining the desired distance between joint pairs.
We demonstrate that the distance between joint pairs for human-wise interactions can be generated using an off-the-shelf Large Language Model.
arXiv Detail & Related papers (2023-11-27T14:32:33Z) - SynH2R: Synthesizing Hand-Object Motions for Learning Human-to-Robot
Handovers [37.49601724575655]
Vision-based human-to-robot handover is an important and challenging task in human-robot interaction.
We introduce a framework that can generate plausible human grasping motions suitable for training the robot.
This allows us to generate synthetic training and testing data with 100x more objects than previous work.
arXiv Detail & Related papers (2023-11-09T18:57:02Z) - Object Motion Guided Human Motion Synthesis [22.08240141115053]
We study the problem of full-body human motion synthesis for the manipulation of large-sized objects.
We propose Object MOtion guided human MOtion synthesis (OMOMO), a conditional diffusion framework.
We develop a novel system that captures full-body human manipulation motions by simply attaching a smartphone to the object being manipulated.
arXiv Detail & Related papers (2023-09-28T08:22:00Z) - DisCo: Disentangled Control for Realistic Human Dance Generation [125.85046815185866]
We introduce DISCO, which includes a novel model architecture with disentangled control to improve the compositionality of dance synthesis.
DisCc can generate high-quality human dance images and videos with diverse appearances and flexible motions.
arXiv Detail & Related papers (2023-06-30T17:37:48Z) - Modelling Human Visual Motion Processing with Trainable Motion Energy
Sensing and a Self-attention Network [1.9458156037869137]
We propose an image-computable model of human motion perception by bridging the gap between biological and computer vision models.
This model architecture aims to capture the computations in V1-MT, the core structure for motion perception in the biological visual system.
In silico neurophysiology reveals that our model's unit responses are similar to mammalian neural recordings regarding motion pooling and speed tuning.
arXiv Detail & Related papers (2023-05-16T04:16:07Z) - Deep state-space modeling for explainable representation, analysis, and
generation of professional human poses [0.0]
This paper introduces three novel methods for creating explainable representations of human movement.
The trained models are used for the full-body dexterity analysis of expert professionals.
arXiv Detail & Related papers (2023-04-13T08:13:10Z) - Learn to Predict How Humans Manipulate Large-sized Objects from
Interactive Motions [82.90906153293585]
We propose a graph neural network, HO-GCN, to fuse motion data and dynamic descriptors for the prediction task.
We show the proposed network that consumes dynamic descriptors can achieve state-of-the-art prediction results and help the network better generalize to unseen objects.
arXiv Detail & Related papers (2022-06-25T09:55:39Z) - GIMO: Gaze-Informed Human Motion Prediction in Context [75.52839760700833]
We propose a large-scale human motion dataset that delivers high-quality body pose sequences, scene scans, and ego-centric views with eye gaze.
Our data collection is not tied to specific scenes, which further boosts the motion dynamics observed from our subjects.
To realize the full potential of gaze, we propose a novel network architecture that enables bidirectional communication between the gaze and motion branches.
arXiv Detail & Related papers (2022-04-20T13:17:39Z) - A Review of Deep Learning Techniques for Markerless Human Motion on
Synthetic Datasets [0.0]
Estimating human posture has recently gained increasing attention in the computer vision community.
We present a model that can predict the skeleton of an animation based solely on 2D images.
The implementation process uses DeepLabCut on its own dataset to perform many necessary steps.
arXiv Detail & Related papers (2022-01-07T15:42:50Z) - HSPACE: Synthetic Parametric Humans Animated in Complex Environments [67.8628917474705]
We build a large-scale photo-realistic dataset, Human-SPACE, of animated humans placed in complex indoor and outdoor environments.
We combine a hundred diverse individuals of varying ages, gender, proportions, and ethnicity, with hundreds of motions and scenes, in order to generate an initial dataset of over 1 million frames.
Assets are generated automatically, at scale, and are compatible with existing real time rendering and game engines.
arXiv Detail & Related papers (2021-12-23T22:27:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.