SGCE-Font: Skeleton Guided Channel Expansion for Chinese Font Generation
- URL: http://arxiv.org/abs/2211.14475v1
- Date: Sat, 26 Nov 2022 04:21:46 GMT
- Title: SGCE-Font: Skeleton Guided Channel Expansion for Chinese Font Generation
- Authors: Jie Zhou, Yefei Wang, Yiyang Yuan, Qing Huang, Jinshan Zeng
- Abstract summary: This paper proposes a novel information guidance module called the skeleton guided channel expansion (SGCE) module for the Chinese font generation.
Numerical results show that the mode collapse issue suffered by the known CycleGAN can be effectively alleviated by equipping with the proposed SGCE module.
- Score: 19.20334101519465
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The automatic generation of Chinese fonts is an important problem involved in
many applications. The predominated methods for the Chinese font generation are
based on the deep generative models, especially the generative adversarial
networks (GANs). However, existing GAN-based methods (say, CycleGAN) for the
Chinese font generation usually suffer from the mode collapse issue, mainly due
to the lack of effective guidance information. This paper proposes a novel
information guidance module called the skeleton guided channel expansion (SGCE)
module for the Chinese font generation through integrating the skeleton
information into the generator with the channel expansion way, motivated by the
observation that the skeleton embodies both local and global structure
information of Chinese characters. We conduct extensive experiments to show the
effectiveness of the proposed module. Numerical results show that the mode
collapse issue suffered by the known CycleGAN can be effectively alleviated by
equipping with the proposed SGCE module, and the CycleGAN equipped with SGCE
outperforms the state-of-the-art models in terms of four important evaluation
metrics and visualization quality. Besides CycleGAN, we also show that the
suggested SGCE module can be adapted to other models for Chinese font
generation as a plug-and-play module to further improve their performance.
Related papers
- A Review of Modern Recommender Systems Using Generative Models (Gen-RecSys) [57.30228361181045]
This survey connects key advancements in recommender systems using Generative Models (Gen-RecSys)
It covers: interaction-driven generative models; the use of large language models (LLM) and textual data for natural language recommendation; and the integration of multimodal models for generating and processing images/videos in RS.
Our work highlights necessary paradigms for evaluating the impact and harm of Gen-RecSys and identifies open challenges.
arXiv Detail & Related papers (2024-03-31T06:57:57Z) - Contextualization Distillation from Large Language Model for Knowledge
Graph Completion [51.126166442122546]
We introduce the Contextualization Distillation strategy, a plug-in-and-play approach compatible with both discriminative and generative KGC frameworks.
Our method begins by instructing large language models to transform compact, structural triplets into context-rich segments.
Comprehensive evaluations across diverse datasets and KGC techniques highlight the efficacy and adaptability of our approach.
arXiv Detail & Related papers (2024-01-28T08:56:49Z) - DeepCalliFont: Few-shot Chinese Calligraphy Font Synthesis by
Integrating Dual-modality Generative Models [20.76773399161289]
Few-shot font generation, especially for Chinese calligraphy fonts, is a challenging and ongoing problem.
We propose a novel model, DeepCalliFont, for few-shot Chinese calligraphy font synthesis by integrating dual-modality generative models.
arXiv Detail & Related papers (2023-12-16T04:23:12Z) - PAI-Diffusion: Constructing and Serving a Family of Open Chinese
Diffusion Models for Text-to-image Synthesis on the Cloud [54.046884854230555]
This paper introduces PAI-Diffusion, a comprehensive framework for Chinese text-to-image synthesis.
It incorporates both general and domain-specific Chinese diffusion models, enabling the generation of contextually relevant images.
It seamlessly integrates with Alibaba Cloud's Machine Learning Platform for AI, providing accessible and scalable solutions.
arXiv Detail & Related papers (2023-09-11T15:18:28Z) - VQ-Font: Few-Shot Font Generation with Structure-Aware Enhancement and
Quantization [52.870638830417]
We propose a VQGAN-based framework (i.e., VQ-Font) to enhance glyph fidelity through token prior refinement and structure-aware enhancement.
Specifically, we pre-train a VQGAN to encapsulate font token prior within a codebook. Subsequently, VQ-Font refines the synthesized glyphs with the codebook to eliminate the domain gap between synthesized and real-world strokes.
arXiv Detail & Related papers (2023-08-27T06:32:20Z) - Chinese Financial Text Emotion Mining: GCGTS -- A Character
Relationship-based Approach for Simultaneous Aspect-Opinion Pair Extraction [7.484918031250864]
Aspect-Opinion Pair Extraction (AOPE) from Chinese financial texts is a specialized task in fine-grained text sentiment analysis.
Previous studies have mainly focused on developing grid annotation schemes within grid-based models to facilitate this extraction process.
We propose a novel method called Graph-based Character-level Grid Tagging Scheme (GCGTS)
The GCGTS method explicitly incorporates syntactic structure using Graph Convolutional Networks (GCN) and unifies the encoding of characters within the same semantic unit (Chinese word level)
arXiv Detail & Related papers (2023-08-04T02:20:56Z) - Diff-Font: Diffusion Model for Robust One-Shot Font Generation [110.45944936952309]
We propose a novel one-shot font generation method based on a diffusion model, named Diff-Font.
The proposed model aims to generate the entire font library by giving only one sample as the reference.
The well-trained Diff-Font is not only robust to font gap and font variation, but also achieved promising performance on difficult character generation.
arXiv Detail & Related papers (2022-12-12T13:51:50Z) - StrokeGAN+: Few-Shot Semi-Supervised Chinese Font Generation with Stroke
Encoding [23.886977380061662]
This paper proposes an effective model called textitGAN+Stroke, which incorporates the stroke encoding and the few-shot semi-supervised scheme into the CycleGAN model.
Experimental results show that the mode collapse issue can be effectively alleviated by the introduced one-bit stroke encoding and few-shot semi-supervised training scheme.
arXiv Detail & Related papers (2022-11-11T13:39:26Z) - GRETEL: Graph Contrastive Topic Enhanced Language Model for Long
Document Extractive Summarization [22.053942327552583]
We propose a graph contrastive topic enhanced language model (GRETEL) for capturing global semantic information.
GRETEL integrates the hierarchical transformer encoder and the graph contrastive learning to fuse the semantic information from the global document context and the gold summary.
Experimental results on both general domain and biomedical datasets demonstrate that our proposed method outperforms SOTA methods.
arXiv Detail & Related papers (2022-08-21T23:09:29Z) - Lattice-BERT: Leveraging Multi-Granularity Representations in Chinese
Pre-trained Language Models [62.41139712595334]
We propose a novel pre-training paradigm for Chinese -- Lattice-BERT.
We construct a lattice graph from the characters and words in a sentence and feed all these text units into transformers.
We show that our model can bring an average increase of 1.5% under the 12-layer setting.
arXiv Detail & Related papers (2021-04-15T02:36:49Z) - StrokeGAN: Reducing Mode Collapse in Chinese Font Generation via Stroke
Encoding [20.877391644999534]
We introduce a one-bit stroke encoding to capture the key mode information of Chinese characters.
We incorporate this mode information into CycleGAN, a popular deep generative model for Chinese font generation.
StrokeGAN is generally outperforms the state-of-the-art methods in terms of content and recognition accuracies.
arXiv Detail & Related papers (2020-12-16T01:36:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.