Enhanced Anime Image Generation Using USE-CMHSA-GAN
- URL: http://arxiv.org/abs/2411.11179v1
- Date: Sun, 17 Nov 2024 21:25:24 GMT
- Title: Enhanced Anime Image Generation Using USE-CMHSA-GAN
- Authors: J. Lu,
- Abstract summary: This paper introduces a novel Generative Adversarial Network model, USE-CMHSA-GAN, designed to produce high-quality anime character images.
Experiments were conducted on the anime-face-dataset, and the results demonstrate that USE-CMHSA-GAN outperforms other benchmark models.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: With the growing popularity of ACG (Anime, Comics, and Games) culture, generating high-quality anime character images has become an important research topic. This paper introduces a novel Generative Adversarial Network model, USE-CMHSA-GAN, designed to produce high-quality anime character images. The model builds upon the traditional DCGAN framework, incorporating USE and CMHSA modules to enhance feature extraction capabilities for anime character images. Experiments were conducted on the anime-face-dataset, and the results demonstrate that USE-CMHSA-GAN outperforms other benchmark models, including DCGAN, VAE-GAN, and WGAN, in terms of FID and IS scores, indicating superior image quality. These findings suggest that USE-CMHSA-GAN is highly effective for anime character image generation and provides new insights for further improving the quality of generative models.
Related papers
- NijiGAN: Transform What You See into Anime with Contrastive Semi-Supervised Learning and Neural Ordinary Differential Equations [0.0]
This research introduces NijiGAN, a novel model for image-to-image translation.
It transforms real-world scenes into high fidelity anime visuals using half of Scenimefy's parameters.
It employs pseudo-paired data generated through Scenimefy for supervised training, eliminating dependence on low-quality paired data.
arXiv Detail & Related papers (2024-12-27T04:46:44Z) - The Dawn of KAN in Image-to-Image (I2I) Translation: Integrating Kolmogorov-Arnold Networks with GANs for Unpaired I2I Translation [0.0]
Kolmogorov-Arnold Network (KAN) can effectively replace the Multi-layer Perceptron (MLP) method in generative AI.
Work suggests KAN could be a valuable component in the broader generative AI domain.
arXiv Detail & Related papers (2024-08-15T15:26:12Z) - RenAIssance: A Survey into AI Text-to-Image Generation in the Era of
Large Model [93.8067369210696]
Text-to-image generation (TTI) refers to the usage of models that could process text input and generate high fidelity images based on text descriptions.
Diffusion models are one prominent type of generative model used for the generation of images through the systematic introduction of noises with repeating steps.
In the era of large models, scaling up model size and the integration with large language models have further improved the performance of TTI models.
arXiv Detail & Related papers (2023-09-02T03:27:20Z) - StyleAvatar3D: Leveraging Image-Text Diffusion Models for High-Fidelity
3D Avatar Generation [103.88928334431786]
We present a novel method for generating high-quality, stylized 3D avatars.
We use pre-trained image-text diffusion models for data generation and a Generative Adversarial Network (GAN)-based 3D generation network for training.
Our approach demonstrates superior performance over current state-of-the-art methods in terms of visual quality and diversity of the produced avatars.
arXiv Detail & Related papers (2023-05-30T13:09:21Z) - Generative Adversarial Networks for Image Super-Resolution: A Survey [101.39605080291783]
Single image super-resolution (SISR) has played an important role in the field of image processing.
Recent generative adversarial networks (GANs) can achieve excellent results on low-resolution images with small samples.
In this paper, we conduct a comparative study of GANs from different perspectives.
arXiv Detail & Related papers (2022-04-28T16:35:04Z) - PetsGAN: Rethinking Priors for Single Image Generation [29.213169685303495]
SinGAN builds a pyramid of GANs to progressively learn the internal patch distribution of the single image.
Due to the lack of high-level information, SinGAN cannot handle the object images well.
Our method gets rid of the time-consuming progressive training scheme and can be trained end-to-end.
arXiv Detail & Related papers (2022-03-03T02:31:50Z) - GIU-GANs: Global Information Utilization for Generative Adversarial
Networks [3.3945834638760948]
In this paper, we propose a new GANs called Involution Generative Adversarial Networks (GIU-GANs)
GIU-GANs leverages a brand new module called the Global Information Utilization (GIU) module, which integrates Squeeze-and-Excitation Networks (SENet) and involution.
Batch Normalization(BN) inevitably ignores the representation differences among noise sampled by the generator, and thus degrades the generated image quality.
arXiv Detail & Related papers (2022-01-25T17:17:15Z) - A Generic Approach for Enhancing GANs by Regularized Latent Optimization [79.00740660219256]
We introduce a generic framework called em generative-model inference that is capable of enhancing pre-trained GANs effectively and seamlessly.
Our basic idea is to efficiently infer the optimal latent distribution for the given requirements using Wasserstein gradient flow techniques.
arXiv Detail & Related papers (2021-12-07T05:22:50Z) - Enhancement of Anime Imaging Enlargement using Modified Super-Resolution
CNN [0.0]
We propose a model based on convolutional neural networks to extract outstanding features of images, enlarge those images, and enhance the quality of Anime images.
The experimental results indicated that our model successfully enhanced the image quality with a larger image-size when compared with the common existing image enlargement and the original SRCNN method.
arXiv Detail & Related papers (2021-10-05T19:38:50Z) - CAGAN: Text-To-Image Generation with Combined Attention GANs [70.3497683558609]
We propose the Combined Attention Generative Adversarial Network (CAGAN) to generate photo-realistic images according to textual descriptions.
The proposed CAGAN uses two attention models: word attention to draw different sub-regions conditioned on related words; and squeeze-and-excitation attention to capture non-linear interaction among channels.
With spectral normalisation to stabilise training, our proposed CAGAN improves the state of the art on the IS and FID on the CUB dataset and the FID on the more challenging COCO dataset.
arXiv Detail & Related papers (2021-04-26T15:46:40Z) - AniGAN: Style-Guided Generative Adversarial Networks for Unsupervised
Anime Face Generation [84.52819242283852]
We propose a novel framework to translate a portrait photo-face into an anime appearance.
Our aim is to synthesize anime-faces which are style-consistent with a given reference anime-face.
Existing methods often fail to transfer the styles of reference anime-faces, or introduce noticeable artifacts/distortions in the local shapes of their generated faces.
arXiv Detail & Related papers (2021-02-24T22:47:38Z) - Anime-to-Real Clothing: Cosplay Costume Generation via Image-to-Image
Translation [2.4660652494309936]
This paper presents an automatic costume image generation method based on image-to-image translation.
We present a novel architecture for generative adversarial networks (GANs) to facilitate high-quality cosplay image generation.
Experiments demonstrated that, with two types of evaluation metrics, the proposed GAN achieves better performance than existing methods.
arXiv Detail & Related papers (2020-08-26T10:34:46Z) - Image Augmentations for GAN Training [57.65145659417266]
We provide insights and guidelines on how to augment images for both vanilla GANs and GANs with regularizations.
Surprisingly, we find that vanilla GANs attain generation quality on par with recent state-of-the-art results.
arXiv Detail & Related papers (2020-06-04T00:16:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.