Improving the Perceptual Quality of 2D Animation Interpolation
- URL: http://arxiv.org/abs/2111.12792v1
- Date: Wed, 24 Nov 2021 20:51:29 GMT
- Title: Improving the Perceptual Quality of 2D Animation Interpolation
- Authors: Shuhong Chen, Matthias Zwicker
- Abstract summary: Traditional 2D animation is labor-intensive, often requiring animators to draw twelve illustrations per second of movement.
Lower framerates result in larger displacements and occlusions, discrete perceptual elements (e.g. lines and solid-color regions) pose difficulties for texture-oriented convolutional networks.
Previous work tried addressing these issues, but used unscalable methods and focused on pixel-perfect performance.
We build a scalable system more appropriately centered on perceptual quality for this artistic domain.
- Score: 37.04208600867858
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Traditional 2D animation is labor-intensive, often requiring animators to
manually draw twelve illustrations per second of movement. While automatic
frame interpolation may ease this burden, the artistic effects inherent to 2D
animation make video synthesis particularly challenging compared to in the
photorealistic domain. Lower framerates result in larger displacements and
occlusions, discrete perceptual elements (e.g. lines and solid-color regions)
pose difficulties for texture-oriented convolutional networks, and exaggerated
nonlinear movements hinder training data collection. Previous work tried
addressing these issues, but used unscalable methods and focused on
pixel-perfect performance. In contrast, we build a scalable system more
appropriately centered on perceptual quality for this artistic domain. Firstly,
we propose a lightweight architecture with a simple yet effective
occlusion-inpainting technique to improve convergence on perceptual metrics
with fewer trainable parameters. Secondly, we design a novel auxiliary module
that leverages the Euclidean distance transform to improve the preservation of
key line and region structures. Thirdly, we automatically double the existing
manually-collected dataset for this task by quantitatively filtering out
movement nonlinearities, allowing us to improve model generalization. Finally,
we establish LPIPS and chamfer distance as strongly preferable to PSNR and SSIM
through a user study, validating our system's emphasis on perceptual quality in
the 2D animation domain.
Related papers
- Thin-Plate Spline-based Interpolation for Animation Line Inbetweening [54.69811179222127]
Chamfer Distance (CD) is commonly adopted for evaluating inbetweening performance.
We propose a simple yet effective method for animation line inbetweening that adopts thin-plate spline-based transformation.
Our method outperforms existing approaches by delivering high-quality results with enhanced fluidity.
arXiv Detail & Related papers (2024-08-17T08:05:31Z) - An Animation-based Augmentation Approach for Action Recognition from Discontinuous Video [11.293897932762809]
Action recognition, an essential component of computer vision, plays a pivotal role in multiple applications.
CNNs suffer performance declines when trained with discontinuous video frames, which is a frequent scenario in real-world settings.
To overcome this issue, we introduce the 4A pipeline, which employs a series of sophisticated techniques.
arXiv Detail & Related papers (2024-04-10T04:59:51Z) - Bridging the Gap: Sketch-Aware Interpolation Network for High-Quality Animation Sketch Inbetweening [58.09847349781176]
We propose a novel deep learning method - Sketch-Aware Interpolation Network (SAIN)
This approach incorporates multi-level guidance that formulates region-level correspondence, stroke-level correspondence and pixel-level dynamics.
A multi-stream U-Transformer is then devised to characterize sketch inbetweening patterns using these multi-level guides through the integration of self / cross-attention mechanisms.
arXiv Detail & Related papers (2023-08-25T09:51:03Z) - RiCS: A 2D Self-Occlusion Map for Harmonizing Volumetric Objects [68.85305626324694]
Ray-marching in Camera Space (RiCS) is a new method to represent the self-occlusions of foreground objects in 3D into a 2D self-occlusion map.
We show that our representation map not only allows us to enhance the image quality but also to model temporally coherent complex shadow effects.
arXiv Detail & Related papers (2022-05-14T05:35:35Z) - Image2Gif: Generating Continuous Realistic Animations with Warping NODEs [0.8218964199015377]
We propose a new framework, Warping Neural ODE, for generating a smooth animation (video frame) in a continuous manner.
This allows us to achieve the smoothness and the realism of an animation with infinitely small time steps between the frames.
We show the application of our work in generating an animation given two frames, in different training settings, including Generative Adversarial Network (GAN) and with $L$ loss.
arXiv Detail & Related papers (2022-05-09T18:39:47Z) - Decoupled Spatial-Temporal Transformer for Video Inpainting [77.8621673355983]
Video aims to fill the given holes with realistic appearance but is still a challenging task even with prosperous deep learning approaches.
Recent works introduce the promising Transformer architecture into deep video inpainting and achieve better performance.
We propose a Decoupled Spatial-Temporal Transformer (DSTT) for improving video inpainting with exceptional efficiency.
arXiv Detail & Related papers (2021-04-14T05:47:46Z) - Deep Animation Video Interpolation in the Wild [115.24454577119432]
In this work, we formally define and study the animation video code problem for the first time.
We propose an effective framework, AnimeInterp, with two dedicated modules in a coarse-to-fine manner.
Notably, AnimeInterp shows favorable perceptual quality and robustness for animation scenarios in the wild.
arXiv Detail & Related papers (2021-04-06T13:26:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.