Unpaired Motion Style Transfer from Video to Animation
- URL: http://arxiv.org/abs/2005.05751v1
- Date: Tue, 12 May 2020 13:21:27 GMT
- Title: Unpaired Motion Style Transfer from Video to Animation
- Authors: Kfir Aberman, Yijia Weng, Dani Lischinski, Daniel Cohen-Or, Baoquan
Chen
- Abstract summary: Transferring the motion style from one animation clip to another, while preserving the motion content of the latter, has been a long-standing problem in character animation.
We present a novel data-driven framework for motion style transfer, which learns from an unpaired collection of motions with style labels.
Our framework is able to extract motion styles directly from videos, bypassing 3D reconstruction, and apply them to the 3D input motion.
- Score: 74.15550388701833
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Transferring the motion style from one animation clip to another, while
preserving the motion content of the latter, has been a long-standing problem
in character animation. Most existing data-driven approaches are supervised and
rely on paired data, where motions with the same content are performed in
different styles. In addition, these approaches are limited to transfer of
styles that were seen during training. In this paper, we present a novel
data-driven framework for motion style transfer, which learns from an unpaired
collection of motions with style labels, and enables transferring motion styles
not observed during training. Furthermore, our framework is able to extract
motion styles directly from videos, bypassing 3D reconstruction, and apply them
to the 3D input motion. Our style transfer network encodes motions into two
latent codes, for content and for style, each of which plays a different role
in the decoding (synthesis) process. While the content code is decoded into the
output motion by several temporal convolutional layers, the style code modifies
deep features via temporally invariant adaptive instance normalization (AdaIN).
Moreover, while the content code is encoded from 3D joint rotations, we learn a
common embedding for style from either 3D or 2D joint positions, enabling style
extraction from videos. Our results are comparable to the state-of-the-art,
despite not requiring paired training data, and outperform other methods when
transferring previously unseen styles. To our knowledge, we are the first to
demonstrate style transfer directly from videos to 3D animations - an ability
which enables one to extend the set of style examples far beyond motions
captured by MoCap systems.
Related papers
- Zero-shot High-fidelity and Pose-controllable Character Animation [89.74818983864832]
Image-to-video (I2V) generation aims to create a video sequence from a single image.
Existing approaches suffer from inconsistency of character appearances and poor preservation of fine details.
We propose PoseAnimate, a novel zero-shot I2V framework for character animation.
arXiv Detail & Related papers (2024-04-21T14:43:31Z) - AniClipart: Clipart Animation with Text-to-Video Priors [28.76809141136148]
We introduce AniClipart, a system that transforms static images into high-quality motion sequences guided by text-to-video priors.
Experimental results show that the proposed AniClipart consistently outperforms existing image-to-video generation models.
arXiv Detail & Related papers (2024-04-18T17:24:28Z) - MoST: Motion Style Transformer between Diverse Action Contents [23.62426940733713]
We propose a novel motion style transformer that effectively disentangles style from content and generates a plausible motion with transferred style from a source motion.
Our method outperforms existing methods and demonstrates exceptionally high quality, particularly in motion pairs with different contents, without the need for post-processing.
arXiv Detail & Related papers (2024-03-10T14:11:25Z) - WAIT: Feature Warping for Animation to Illustration video Translation
using GANs [12.681919619814419]
We introduce a new problem for video stylizing where an unordered set of images are used.
Most of the video-to-video translation methods are built on an image-to-image translation model.
We propose a new generator network with feature warping layers which overcomes the limitations of the previous methods.
arXiv Detail & Related papers (2023-10-07T19:45:24Z) - MotionBERT: A Unified Perspective on Learning Human Motion
Representations [46.67364057245364]
We present a unified perspective on tackling various human-centric video tasks by learning human motion representations from large-scale and heterogeneous data resources.
We propose a pretraining stage in which a motion encoder is trained to recover the underlying 3D motion from noisy partial 2D observations.
We implement motion encoder with a Dual-stream Spatio-temporal Transformer (DSTformer) neural network.
arXiv Detail & Related papers (2022-10-12T19:46:25Z) - Self-Supervised Equivariant Scene Synthesis from Video [84.15595573718925]
We propose a framework to learn scene representations from video that are automatically delineated into background, characters, and animations.
After training, we can manipulate image encodings in real time to create unseen combinations of the delineated components.
We demonstrate results on three datasets: Moving MNIST with backgrounds, 2D video game sprites, and Fashion Modeling.
arXiv Detail & Related papers (2021-02-01T14:17:31Z) - Animating Pictures with Eulerian Motion Fields [90.30598913855216]
We show a fully automatic method for converting a still image into a realistic animated looping video.
We target scenes with continuous fluid motion, such as flowing water and billowing smoke.
We propose a novel video looping technique that flows features both forward and backward in time and then blends the results.
arXiv Detail & Related papers (2020-11-30T18:59:06Z) - 3DSNet: Unsupervised Shape-to-Shape 3D Style Transfer [66.48720190245616]
We propose a learning-based approach for style transfer between 3D objects.
The proposed method can synthesize new 3D shapes both in the form of point clouds and meshes.
We extend our technique to implicitly learn the multimodal style distribution of the chosen domains.
arXiv Detail & Related papers (2020-11-26T16:59:12Z) - First Order Motion Model for Image Animation [90.712718329677]
Image animation consists of generating a video sequence so that an object in a source image is animated according to the motion of a driving video.
Our framework addresses this problem without using any annotation or prior information about the specific object to animate.
arXiv Detail & Related papers (2020-02-29T07:08:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.