Interactive Video Stylization Using Few-Shot Patch-Based Training
- URL: http://arxiv.org/abs/2004.14489v1
- Date: Wed, 29 Apr 2020 21:33:28 GMT
- Title: Interactive Video Stylization Using Few-Shot Patch-Based Training
- Authors: Ond\v{r}ej Texler, David Futschik, Michal Ku\v{c}era, Ond\v{r}ej
Jamri\v{s}ka, \v{S}\'arka Sochorov\'a, Menglei Chai, Sergey Tulyakov, and
Daniel S\'ykora
- Abstract summary: We present a learning-based method to propagate the style from a few selecteds to the rest of the sequence.
The resulting stylization is semantically meaningful, i.e., specific parts of moving objects are stylized according to the artist's intention.
We demonstrate its practical utility in various interactive scenarios, where the user paints over a selected appearance and sees her style transferred to an existing recorded sequence or a live video stream.
- Score: 23.801384381295698
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we present a learning-based method to the keyframe-based video
stylization that allows an artist to propagate the style from a few selected
keyframes to the rest of the sequence. Its key advantage is that the resulting
stylization is semantically meaningful, i.e., specific parts of moving objects
are stylized according to the artist's intention. In contrast to previous style
transfer techniques, our approach does not require any lengthy pre-training
process nor a large training dataset. We demonstrate how to train an appearance
translation network from scratch using only a few stylized exemplars while
implicitly preserving temporal consistency. This leads to a video stylization
framework that supports real-time inference, parallel processing, and random
access to an arbitrary output frame. It can also merge the content from
multiple keyframes without the need to perform an explicit blending operation.
We demonstrate its practical utility in various interactive scenarios, where
the user paints over a selected keyframe and sees her style transferred to an
existing recorded sequence or a live video stream.
Related papers
- Framer: Interactive Frame Interpolation [73.06734414930227]
Framer targets producing smoothly transitioning frames between two images as per user creativity.
Our approach supports customizing the transition process by tailoring the trajectory of some selected keypoints.
It is noteworthy that our system also offers an "autopilot" mode, where we introduce a module to estimate the keypoints and the trajectory automatically.
arXiv Detail & Related papers (2024-10-24T17:59:51Z) - Learning Inclusion Matching for Animation Paint Bucket Colorization [76.4507878427755]
We introduce a new learning-based inclusion matching pipeline, which directs the network to comprehend the inclusion relationships between segments.
Our method features a two-stage pipeline that integrates a coarse color warping module with an inclusion matching module.
To facilitate the training of our network, we also develope a unique dataset, referred to as PaintBucket-Character.
arXiv Detail & Related papers (2024-03-27T08:32:48Z) - Sketch Video Synthesis [52.134906766625164]
We propose a novel framework for sketching videos represented by the frame-wise B'ezier curve.
Our method unlocks applications in sketch-based video editing and video doodling, enabled through video composition.
arXiv Detail & Related papers (2023-11-26T14:14:04Z) - WAIT: Feature Warping for Animation to Illustration video Translation
using GANs [12.681919619814419]
We introduce a new problem for video stylizing where an unordered set of images are used.
Most of the video-to-video translation methods are built on an image-to-image translation model.
We propose a new generator network with feature warping layers which overcomes the limitations of the previous methods.
arXiv Detail & Related papers (2023-10-07T19:45:24Z) - Rerender A Video: Zero-Shot Text-Guided Video-to-Video Translation [93.18163456287164]
This paper proposes a novel text-guided video-to-video translation framework to adapt image models to videos.
Our framework achieves global style and local texture temporal consistency at a low cost.
arXiv Detail & Related papers (2023-06-13T17:52:23Z) - Style-A-Video: Agile Diffusion for Arbitrary Text-based Video Style
Transfer [13.098901971644656]
This paper proposes a zero-shot video stylization method named Style-A-Video.
Uses a generative pre-trained transformer with an image latent diffusion model to achieve a concise text-controlled video stylization.
Tests show that we can attain superior content preservation and stylistic performance while incurring less consumption than previous solutions.
arXiv Detail & Related papers (2023-05-09T14:03:27Z) - A Unified Arbitrary Style Transfer Framework via Adaptive Contrastive
Learning [84.8813842101747]
Unified Contrastive Arbitrary Style Transfer (UCAST) is a novel style representation learning and transfer framework.
We present an adaptive contrastive learning scheme for style transfer by introducing an input-dependent temperature.
Our framework consists of three key components, i.e., a parallel contrastive learning scheme for style representation and style transfer, a domain enhancement module for effective learning of style distribution, and a generative network for style transfer.
arXiv Detail & Related papers (2023-03-09T04:35:00Z) - STALP: Style Transfer with Auxiliary Limited Pairing [36.23393954839379]
We present an approach to example-based stylization of images that uses a single pair of a source image and its stylized counterpart.
We demonstrate how to train an image translation network that can perform real-time semantically meaningful style transfer to a set of target images.
arXiv Detail & Related papers (2021-10-20T11:38:41Z) - Real-time Localized Photorealistic Video Style Transfer [25.91181753178577]
We present a novel algorithm for transferring artistic styles of semantically meaningful local regions of an image onto local regions of a target video.
Our method, based on a deep neural network architecture inspired by recent work in photorealistic style transfer, is real-time and works on arbitrary inputs.
We demonstrate our method on a variety of style images and target videos, including the ability to transfer different styles onto multiple objects simultaneously.
arXiv Detail & Related papers (2020-10-20T06:21:09Z) - Deep Sketch-guided Cartoon Video Inbetweening [24.00033622396297]
We propose a framework to produce cartoon videos by fetching the color information from two inputs while following the animated motion guided by a user sketch.
By explicitly considering the correspondence between frames and the sketch, we can achieve higher quality results than other image synthesis methods.
arXiv Detail & Related papers (2020-08-10T14:22:04Z) - Unpaired Motion Style Transfer from Video to Animation [74.15550388701833]
Transferring the motion style from one animation clip to another, while preserving the motion content of the latter, has been a long-standing problem in character animation.
We present a novel data-driven framework for motion style transfer, which learns from an unpaired collection of motions with style labels.
Our framework is able to extract motion styles directly from videos, bypassing 3D reconstruction, and apply them to the 3D input motion.
arXiv Detail & Related papers (2020-05-12T13:21:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.