DG-Font: Deformable Generative Networks for Unsupervised Font Generation
- URL: http://arxiv.org/abs/2104.03064v2
- Date: Thu, 8 Apr 2021 14:19:26 GMT
- Title: DG-Font: Deformable Generative Networks for Unsupervised Font Generation
- Authors: Yangchen Xie and Xinyuan Chen and Li Sun and Yue Lu
- Abstract summary: We propose novel deformable generative networks for unsupervised font generation (DGFont)
We introduce a feature deformation skip connection (FDSC) which predicts pairs of displacement maps and employs the predicted maps to apply deformable convolution to the low-level feature maps from the content encoder.
Experiments demonstrate that our model generates characters in higher quality than state-of-art methods.
- Score: 14.178381391124036
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Font generation is a challenging problem especially for some writing systems
that consist of a large number of characters and has attracted a lot of
attention in recent years. However, existing methods for font generation are
often in supervised learning. They require a large number of paired data, which
is labor-intensive and expensive to collect. Besides, common image-to-image
translation models often define style as the set of textures and colors, which
cannot be directly applied to font generation. To address these problems, we
propose novel deformable generative networks for unsupervised font generation
(DGFont). We introduce a feature deformation skip connection (FDSC) which
predicts pairs of displacement maps and employs the predicted maps to apply
deformable convolution to the low-level feature maps from the content encoder.
The outputs of FDSC are fed into a mixer to generate the final results. Taking
advantage of FDSC, the mixer outputs a high-quality character with a complete
structure. To further improve the quality of generated images, we use three
deformable convolution layers in the content encoder to learn style-invariant
feature representations. Experiments demonstrate that our model generates
characters in higher quality than state-of-art methods. The source code is
available at https://github.com/ecnuycxie/DG-Font.
Related papers
- JoyType: A Robust Design for Multilingual Visual Text Creation [14.441897362967344]
We introduce a novel approach for multilingual visual text creation, named JoyType.
JoyType is designed to maintain the font style of text during the image generation process.
Our evaluations, based on both visual and accuracy metrics, demonstrate that JoyType significantly outperforms existing state-of-the-art methods.
arXiv Detail & Related papers (2024-09-26T04:23:17Z) - FontDiffuser: One-Shot Font Generation via Denoising Diffusion with
Multi-Scale Content Aggregation and Style Contrastive Learning [45.696909070215476]
FontDiffuser is a diffusion-based image-to-image one-shot font generation method.
It consistently excels on complex characters and large style changes compared to previous methods.
arXiv Detail & Related papers (2023-12-19T13:23:20Z) - VQ-Font: Few-Shot Font Generation with Structure-Aware Enhancement and
Quantization [52.870638830417]
We propose a VQGAN-based framework (i.e., VQ-Font) to enhance glyph fidelity through token prior refinement and structure-aware enhancement.
Specifically, we pre-train a VQGAN to encapsulate font token prior within a codebook. Subsequently, VQ-Font refines the synthesized glyphs with the codebook to eliminate the domain gap between synthesized and real-world strokes.
arXiv Detail & Related papers (2023-08-27T06:32:20Z) - CF-Font: Content Fusion for Few-shot Font Generation [63.79915037830131]
We propose a content fusion module (CFM) to project the content feature into a linear space defined by the content features of basis fonts.
Our method also allows to optimize the style representation vector of reference images.
We have evaluated our method on a dataset of 300 fonts with 6.5k characters each.
arXiv Detail & Related papers (2023-03-24T14:18:40Z) - DGFont++: Robust Deformable Generative Networks for Unsupervised Font
Generation [19.473023811252116]
We propose a robust deformable generative network for unsupervised font generation (abbreviated as DGFont++)
To distinguish different styles, we train our model with a multi-task discriminator, which ensures that each style can be discriminated independently.
Experiments demonstrate that our model is able to generate character images of higher quality than state-of-the-art methods.
arXiv Detail & Related papers (2022-12-30T14:35:10Z) - Diff-Font: Diffusion Model for Robust One-Shot Font Generation [110.45944936952309]
We propose a novel one-shot font generation method based on a diffusion model, named Diff-Font.
The proposed model aims to generate the entire font library by giving only one sample as the reference.
The well-trained Diff-Font is not only robust to font gap and font variation, but also achieved promising performance on difficult character generation.
arXiv Detail & Related papers (2022-12-12T13:51:50Z) - Scalable Font Reconstruction with Dual Latent Manifolds [55.29525824849242]
We propose a deep generative model that performs typography analysis and font reconstruction.
Our approach enables us to massively scale up the number of character types we can effectively model.
We evaluate on the task of font reconstruction over various datasets representing character types of many languages.
arXiv Detail & Related papers (2021-09-10T20:37:43Z) - Font Completion and Manipulation by Cycling Between Multi-Modality
Representations [113.26243126754704]
We innovate to explore the generation of font glyphs as 2D graphic objects with the graph as an intermediate representation.
We formulate a cross-modality cycled image-to-image structure with a graph between an image encoder and an image.
Our model generates improved results than both image-to-image baseline and previous state-of-the-art methods for glyph completion.
arXiv Detail & Related papers (2021-08-30T02:43:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.