Enhancement of Anime Imaging Enlargement using Modified Super-Resolution
CNN
- URL: http://arxiv.org/abs/2110.02321v1
- Date: Tue, 5 Oct 2021 19:38:50 GMT
- Title: Enhancement of Anime Imaging Enlargement using Modified Super-Resolution
CNN
- Authors: Tanakit Intaniyom, Warinthorn Thananporn, and Kuntpong Woraratpanya
- Abstract summary: We propose a model based on convolutional neural networks to extract outstanding features of images, enlarge those images, and enhance the quality of Anime images.
The experimental results indicated that our model successfully enhanced the image quality with a larger image-size when compared with the common existing image enlargement and the original SRCNN method.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Anime is a storytelling medium similar to movies and books. Anime images are
a kind of artworks, which are almost entirely drawn by hand. Hence, reproducing
existing Anime with larger sizes and higher quality images is expensive.
Therefore, we proposed a model based on convolutional neural networks to
extract outstanding features of images, enlarge those images, and enhance the
quality of Anime images. We trained the model with a training set of 160 images
and a validation set of 20 images. We tested the trained model with a testing
set of 20 images. The experimental results indicated that our model
successfully enhanced the image quality with a larger image-size when compared
with the common existing image enlargement and the original SRCNN method.
Related papers
- AnimeDL-2M: Million-Scale AI-Generated Anime Image Detection and Localization in Diffusion Era [11.94929097375473]
Misrepresentations of AI-generated images as hand-drawn artwork pose serious threats to the anime community and industry.
We propose AnimeDL-2M, the first large-scale benchmark for anime IMDL with comprehensive annotations.
We also propose AniXplore, a novel model tailored to the visual characteristics of anime imagery.
arXiv Detail & Related papers (2025-04-15T09:41:08Z) - Enhanced Anime Image Generation Using USE-CMHSA-GAN [0.0]
This paper introduces a novel Generative Adversarial Network model, USE-CMHSA-GAN, designed to produce high-quality anime character images.
Experiments were conducted on the anime-face-dataset, and the results demonstrate that USE-CMHSA-GAN outperforms other benchmark models.
arXiv Detail & Related papers (2024-11-17T21:25:24Z) - PrefPaint: Aligning Image Inpainting Diffusion Model with Human Preference [62.72779589895124]
We make the first attempt to align diffusion models for image inpainting with human aesthetic standards via a reinforcement learning framework.
We train a reward model with a dataset we construct, consisting of nearly 51,000 images annotated with human preferences.
Experiments on inpainting comparison and downstream tasks, such as image extension and 3D reconstruction, demonstrate the effectiveness of our approach.
arXiv Detail & Related papers (2024-10-29T11:49:39Z) - Illustrious: an Open Advanced Illustration Model [7.428509329724737]
We develop a text-to-image anime image generative model, called Illustrious, to achieve high resolution, dynamic color range images, and high restoration ability.
We focus on three critical approaches for model improvement. First, we delve into the significance of the batch size and dropout control, which enables faster learning of controllable token based concept activations.
Second, we increase the training resolution of images, affecting the accurate depiction of character anatomy in much higher resolution, extending its generation capability over 20MP with proper methods.
arXiv Detail & Related papers (2024-09-30T04:59:12Z) - Autoregressive Model Beats Diffusion: Llama for Scalable Image Generation [52.509092010267665]
We introduce LlamaGen, a new family of image generation models that apply original next-token prediction'' paradigm of large language models to visual generation domain.
It is an affirmative answer to whether vanilla autoregressive models, e.g., Llama, without inductive biases on visual signals can achieve state-of-the-art image generation performance if scaling properly.
arXiv Detail & Related papers (2024-06-10T17:59:52Z) - Improving face generation quality and prompt following with synthetic captions [57.47448046728439]
We introduce a training-free pipeline designed to generate accurate appearance descriptions from images of people.
We then use these synthetic captions to fine-tune a text-to-image diffusion model.
Our results demonstrate that this approach significantly improves the model's ability to generate high-quality, realistic human faces.
arXiv Detail & Related papers (2024-05-17T15:50:53Z) - xT: Nested Tokenization for Larger Context in Large Images [79.37673340393475]
xT is a framework for vision transformers which aggregates global context with local details.
We are able to increase accuracy by up to 8.6% on challenging classification tasks.
arXiv Detail & Related papers (2024-03-04T10:29:58Z) - APISR: Anime Production Inspired Real-World Anime Super-Resolution [15.501488335115269]
We argue that video networks and datasets are not necessary for anime SR due to the repetition use of hand-drawing frames.
Instead, we propose an anime image collection pipeline by choosing the least compressed and the most informative frames from the video sources.
We evaluate our method through extensive experiments on the public benchmark, showing our method outperforms state-of-the-art anime dataset-trained approaches.
arXiv Detail & Related papers (2024-03-03T19:52:43Z) - Emu: Enhancing Image Generation Models Using Photogenic Needles in a
Haystack [75.00066365801993]
Training text-to-image models with web scale image-text pairs enables the generation of a wide range of visual concepts from text.
These pre-trained models often face challenges when it comes to generating highly aesthetic images.
We propose quality-tuning to guide a pre-trained model to exclusively generate highly visually appealing images.
arXiv Detail & Related papers (2023-09-27T17:30:19Z) - Generative Adversarial Networks for photo to Hayao Miyazaki style
cartoons [0.0]
We trained a Generative Adversial Network(GAN) on over 60 000 images from works by Hayao Miyazaki at Studio Ghibli.
Our model on average outranked state-of-the-art methods on cartoon-likeness.
arXiv Detail & Related papers (2020-05-15T19:26:11Z) - Unsupervised Real Image Super-Resolution via Generative Variational
AutoEncoder [47.53609520395504]
We revisit the classic example based image super-resolution approaches and come up with a novel generative model for perceptual image super-resolution.
We propose a joint image denoising and super-resolution model via Variational AutoEncoder.
With the aid of the discriminator, an additional overhead of super-resolution subnetwork is attached to super-resolve the denoised image with photo-realistic visual quality.
arXiv Detail & Related papers (2020-04-27T13:49:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.