MotionVideoGAN: A Novel Video Generator Based on the Motion Space
Learned from Image Pairs
- URL: http://arxiv.org/abs/2303.02906v1
- Date: Mon, 6 Mar 2023 05:52:13 GMT
- Title: MotionVideoGAN: A Novel Video Generator Based on the Motion Space
Learned from Image Pairs
- Authors: Jingyuan Zhu, Huimin Ma, Jiansheng Chen, and Jian Yuan
- Abstract summary: We present MotionVideoGAN, a novel video generator synthesizing videos based on the motion space learned by pre-trained image pair generators.
Motion codes help us edit images within the motion space since the edited image shares the same contents with the other unchanged one in image pairs.
Our approach achieves state-of-the-art performance on the most complex video dataset ever used for unconditional video generation evaluation, UCF101.
- Score: 16.964371778504297
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Video generation has achieved rapid progress benefiting from high-quality
renderings provided by powerful image generators. We regard the video synthesis
task as generating a sequence of images sharing the same contents but varying
in motions. However, most previous video synthesis frameworks based on
pre-trained image generators treat content and motion generation separately,
leading to unrealistic generated videos. Therefore, we design a novel framework
to build the motion space, aiming to achieve content consistency and fast
convergence for video generation. We present MotionVideoGAN, a novel video
generator synthesizing videos based on the motion space learned by pre-trained
image pair generators. Firstly, we propose an image pair generator named
MotionStyleGAN to generate image pairs sharing the same contents and producing
various motions. Then we manage to acquire motion codes to edit one image in
the generated image pairs and keep the other unchanged. The motion codes help
us edit images within the motion space since the edited image shares the same
contents with the other unchanged one in image pairs. Finally, we introduce a
latent code generator to produce latent code sequences using motion codes for
video generation. Our approach achieves state-of-the-art performance on the
most complex video dataset ever used for unconditional video generation
evaluation, UCF101.
Related papers
- EasyAnimate: A High-Performance Long Video Generation Method based on Transformer Architecture [11.587428534308945]
EasyAnimate is an advanced method for video generation that leverages the power of transformer architecture for high-performance outcomes.
We have expanded the DiT framework originally designed for 2D image synthesis to accommodate the complexities of 3D video generation by incorporating a motion module block.
We provide a holistic ecosystem for video production based on DiT, encompassing aspects such as data pre-processing, VAE training, DiT models training, and end-to-end video inference.
arXiv Detail & Related papers (2024-05-29T11:11:07Z) - GenDeF: Learning Generative Deformation Field for Video Generation [89.49567113452396]
We propose to render a video by warping one static image with a generative deformation field (GenDeF)
Such a pipeline enjoys three appealing advantages.
arXiv Detail & Related papers (2023-12-07T18:59:41Z) - MoVideo: Motion-Aware Video Generation with Diffusion Models [97.03352319694795]
We propose a novel motion-aware generation (MoVideo) framework that takes motion into consideration from two aspects: video depth and optical flow.
MoVideo achieves state-of-the-art results in both text-to-video and image-to-video generation, showing promising prompt consistency, frame consistency and visual quality.
arXiv Detail & Related papers (2023-11-19T13:36:03Z) - One Style is All you Need to Generate a Video [0.9558392439655012]
We introduce a novel temporal generator based on a set of learned sinusoidal bases.
Our method learns dynamic representations of various actions that are independent of image content and can be transferred between different actors.
arXiv Detail & Related papers (2023-10-27T01:17:48Z) - LaMD: Latent Motion Diffusion for Video Generation [69.4111397077229]
latent motion diffusion (LaMD) framework consists of a motion-decomposed video autoencoder and a diffusion-based motion generator.
Results show that LaMD generates high-quality videos with a wide range of motions, from dynamics to highly controllable movements.
arXiv Detail & Related papers (2023-04-23T10:32:32Z) - Text2Video-Zero: Text-to-Image Diffusion Models are Zero-Shot Video
Generators [70.17041424896507]
Recent text-to-video generation approaches rely on computationally heavy training and require large-scale video datasets.
We propose a new task of zero-shot text-to-video generation using existing text-to-image synthesis methods.
Our method performs comparably or sometimes better than recent approaches, despite not being trained on additional video data.
arXiv Detail & Related papers (2023-03-23T17:01:59Z) - Towards Smooth Video Composition [59.134911550142455]
Video generation requires consistent and persistent frames with dynamic content over time.
This work investigates modeling the temporal relations for composing video with arbitrary length, from a few frames to even infinite, using generative adversarial networks (GANs)
We show that the alias-free operation for single image generation, together with adequately pre-learned knowledge, brings a smooth frame transition without compromising the per-frame quality.
arXiv Detail & Related papers (2022-12-14T18:54:13Z) - A Good Image Generator Is What You Need for High-Resolution Video
Synthesis [73.82857768949651]
We present a framework that leverages contemporary image generators to render high-resolution videos.
We frame the video synthesis problem as discovering a trajectory in the latent space of a pre-trained and fixed image generator.
We introduce a motion generator that discovers the desired trajectory, in which content and motion are disentangled.
arXiv Detail & Related papers (2021-04-30T15:38:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.