Generative Adversarial Networks for photo to Hayao Miyazaki style
cartoons
- URL: http://arxiv.org/abs/2005.07702v1
- Date: Fri, 15 May 2020 19:26:11 GMT
- Title: Generative Adversarial Networks for photo to Hayao Miyazaki style
cartoons
- Authors: Filip Andersson, Simon Arvidsson
- Abstract summary: We trained a Generative Adversial Network(GAN) on over 60 000 images from works by Hayao Miyazaki at Studio Ghibli.
Our model on average outranked state-of-the-art methods on cartoon-likeness.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This paper takes on the problem of transferring the style of cartoon images
to real-life photographic images by implementing previous work done by
CartoonGAN. We trained a Generative Adversial Network(GAN) on over 60 000
images from works by Hayao Miyazaki at Studio Ghibli. To evaluate our results,
we conducted a qualitative survey comparing our results with two
state-of-the-art methods. 117 survey results indicated that our model on
average outranked state-of-the-art methods on cartoon-likeness.
Related papers
- Sakuga-42M Dataset: Scaling Up Cartoon Research [4.676528353567339]
Sakuga-42M comprises 42 milliontexts covering various artistic styles, regions, and years, with comprehensive semantic annotations.
Our motivation is to introduce large-scaling to cartoon research and foster generalization and robustness in future cartoon applications.
arXiv Detail & Related papers (2024-05-13T01:50:05Z) - Emu Video: Factorizing Text-to-Video Generation by Explicit Image Conditioning [59.01091079005586]
We present Emu Video, a text-to-video generation model that factorizes the generation into two steps.
We identify critical design decisions--adjusted noise schedules for diffusion, and multi-stage training.
In human evaluations, our generated videos are strongly preferred in quality compared to all prior work.
arXiv Detail & Related papers (2023-11-17T18:59:04Z) - Automatic Animation of Hair Blowing in Still Portrait Photos [61.54919805051212]
We propose a novel approach to animate human hair in a still portrait photo.
Considering the complexity of hair structure, we innovatively treat hair wisp extraction as an instance segmentation problem.
We propose a wisp-aware animation module that animates hair wisps with pleasing motions without noticeable artifacts.
arXiv Detail & Related papers (2023-09-25T15:11:40Z) - Generating Animatable 3D Cartoon Faces from Single Portraits [51.15618892675337]
We present a novel framework to generate animatable 3D cartoon faces from a single portrait image.
We propose a two-stage reconstruction method to recover the 3D cartoon face with detailed texture.
Finally, we propose a semantic preserving face rigging method based on manually created templates and deformation transfer.
arXiv Detail & Related papers (2023-07-04T04:12:50Z) - Sketch2Saliency: Learning to Detect Salient Objects from Human Drawings [99.9788496281408]
We study how sketches can be used as a weak label to detect salient objects present in an image.
To accomplish this, we introduce a photo-to-sketch generation model that aims to generate sequential sketch coordinates corresponding to a given visual photo.
Tests prove our hypothesis and delineate how our sketch-based saliency detection model gives a competitive performance compared to the state-of-the-art.
arXiv Detail & Related papers (2023-03-20T23:46:46Z) - CNNs and Transformers Perceive Hybrid Images Similar to Humans [47.64219291655723]
We show that predictions of deep learning vision models qualitatively matches with the human perception of hybrid images.
Our results provide yet another evidence in support of the hypothesis that Convolutional Neural Networks (CNNs) and Transformers are good at modeling the feedforward sweep of information in the ventral stream of visual cortex.
arXiv Detail & Related papers (2022-03-19T21:37:07Z) - Enhancement of Anime Imaging Enlargement using Modified Super-Resolution
CNN [0.0]
We propose a model based on convolutional neural networks to extract outstanding features of images, enlarge those images, and enhance the quality of Anime images.
The experimental results indicated that our model successfully enhanced the image quality with a larger image-size when compared with the common existing image enlargement and the original SRCNN method.
arXiv Detail & Related papers (2021-10-05T19:38:50Z) - NTIRE 2021 Challenge on Perceptual Image Quality Assessment [128.83256694901726]
This paper reports on the NTIRE 2021 challenge on perceptual image quality assessment (IQA)
It was held in conjunction with the New Trends in Image Restoration and Enhancement workshop (NTIRE) at CVPR 2021.
As a new type of image processing technology, perceptual image processing algorithms based on Generative Adversarial Networks (GAN) have produced images with more realistic textures.
arXiv Detail & Related papers (2021-05-07T05:36:54Z) - toon2real: Translating Cartoon Images to Realistic Images [1.4419517737536707]
We apply several state-of-the-art models to perform this task; however, they fail to perform good quality translations.
We propose a method based on CycleGAN model for image translation from cartoon domain to photo-realistic domain.
We demonstrate our experimental results and show that our proposed model has achieved the lowest Frechet Inception Distance score and better results compared to another state-of-the-art technique, UNIT.
arXiv Detail & Related papers (2021-02-01T20:22:05Z) - Anime-to-Real Clothing: Cosplay Costume Generation via Image-to-Image
Translation [2.4660652494309936]
This paper presents an automatic costume image generation method based on image-to-image translation.
We present a novel architecture for generative adversarial networks (GANs) to facilitate high-quality cosplay image generation.
Experiments demonstrated that, with two types of evaluation metrics, the proposed GAN achieves better performance than existing methods.
arXiv Detail & Related papers (2020-08-26T10:34:46Z) - Words as Art Materials: Generating Paintings with Sequential GANs [8.249180979158815]
We investigate the generation of artistic images on a large variance dataset.
This dataset includes images with variations, for example, in shape, color, and content.
We propose a sequential Generative Adversarial Network model.
arXiv Detail & Related papers (2020-07-08T19:17:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.