DGFont++: Robust Deformable Generative Networks for Unsupervised Font
Generation
- URL: http://arxiv.org/abs/2212.14742v1
- Date: Fri, 30 Dec 2022 14:35:10 GMT
- Title: DGFont++: Robust Deformable Generative Networks for Unsupervised Font
Generation
- Authors: Xinyuan Chen, Yangchen Xie, Li Sun and Yue Lu
- Abstract summary: We propose a robust deformable generative network for unsupervised font generation (abbreviated as DGFont++)
To distinguish different styles, we train our model with a multi-task discriminator, which ensures that each style can be discriminated independently.
Experiments demonstrate that our model is able to generate character images of higher quality than state-of-the-art methods.
- Score: 19.473023811252116
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Automatic font generation without human experts is a practical and
significant problem, especially for some languages that consist of a large
number of characters. Existing methods for font generation are often in
supervised learning. They require a large number of paired data, which are
labor-intensive and expensive to collect. In contrast, common unsupervised
image-to-image translation methods are not applicable to font generation, as
they often define style as the set of textures and colors. In this work, we
propose a robust deformable generative network for unsupervised font generation
(abbreviated as DGFont++). We introduce a feature deformation skip connection
(FDSC) to learn local patterns and geometric transformations between fonts. The
FDSC predicts pairs of displacement maps and employs the predicted maps to
apply deformable convolution to the low-level content feature maps. The outputs
of FDSC are fed into a mixer to generate final results. Moreover, we introduce
contrastive self-supervised learning to learn a robust style representation for
fonts by understanding the similarity and dissimilarities of fonts. To
distinguish different styles, we train our model with a multi-task
discriminator, which ensures that each style can be discriminated
independently. In addition to adversarial loss, another two reconstruction
losses are adopted to constrain the domain-invariant characteristics between
generated images and content images. Taking advantage of FDSC and the adopted
loss functions, our model is able to maintain spatial information and generates
high-quality character images in an unsupervised manner. Experiments
demonstrate that our model is able to generate character images of higher
quality than state-of-the-art methods.
Related papers
- JoyType: A Robust Design for Multilingual Visual Text Creation [14.441897362967344]
We introduce a novel approach for multilingual visual text creation, named JoyType.
JoyType is designed to maintain the font style of text during the image generation process.
Our evaluations, based on both visual and accuracy metrics, demonstrate that JoyType significantly outperforms existing state-of-the-art methods.
arXiv Detail & Related papers (2024-09-26T04:23:17Z) - FontDiffuser: One-Shot Font Generation via Denoising Diffusion with
Multi-Scale Content Aggregation and Style Contrastive Learning [45.696909070215476]
FontDiffuser is a diffusion-based image-to-image one-shot font generation method.
It consistently excels on complex characters and large style changes compared to previous methods.
arXiv Detail & Related papers (2023-12-19T13:23:20Z) - DS-Fusion: Artistic Typography via Discriminated and Stylized Diffusion [10.75789076591325]
We introduce a novel method to automatically generate an artistic typography by stylizing one or more letter fonts.
Our approach utilizes large language models to bridge texts and visual images for stylization and build an unsupervised generative model.
arXiv Detail & Related papers (2023-03-16T19:12:52Z) - Few-shot Font Generation by Learning Style Difference and Similarity [84.76381937516356]
We propose a novel font generation approach by learning the Difference between different styles and the Similarity of the same style (DS-Font)
Specifically, we propose a multi-layer style projector for style encoding and realize a distinctive style representation via our proposed Cluster-level Contrastive Style (CCS) loss.
arXiv Detail & Related papers (2023-01-24T13:57:25Z) - Diff-Font: Diffusion Model for Robust One-Shot Font Generation [110.45944936952309]
We propose a novel one-shot font generation method based on a diffusion model, named Diff-Font.
The proposed model aims to generate the entire font library by giving only one sample as the reference.
The well-trained Diff-Font is not only robust to font gap and font variation, but also achieved promising performance on difficult character generation.
arXiv Detail & Related papers (2022-12-12T13:51:50Z) - Scalable Font Reconstruction with Dual Latent Manifolds [55.29525824849242]
We propose a deep generative model that performs typography analysis and font reconstruction.
Our approach enables us to massively scale up the number of character types we can effectively model.
We evaluate on the task of font reconstruction over various datasets representing character types of many languages.
arXiv Detail & Related papers (2021-09-10T20:37:43Z) - Font Completion and Manipulation by Cycling Between Multi-Modality
Representations [113.26243126754704]
We innovate to explore the generation of font glyphs as 2D graphic objects with the graph as an intermediate representation.
We formulate a cross-modality cycled image-to-image structure with a graph between an image encoder and an image.
Our model generates improved results than both image-to-image baseline and previous state-of-the-art methods for glyph completion.
arXiv Detail & Related papers (2021-08-30T02:43:29Z) - A Multi-Implicit Neural Representation for Fonts [79.6123184198301]
font-specific discontinuities like edges and corners are difficult to represent using neural networks.
We introduce textitmulti-implicits to represent fonts as a permutation-in set of learned implict functions, without losing features.
arXiv Detail & Related papers (2021-06-12T21:40:11Z) - DG-Font: Deformable Generative Networks for Unsupervised Font Generation [14.178381391124036]
We propose novel deformable generative networks for unsupervised font generation (DGFont)
We introduce a feature deformation skip connection (FDSC) which predicts pairs of displacement maps and employs the predicted maps to apply deformable convolution to the low-level feature maps from the content encoder.
Experiments demonstrate that our model generates characters in higher quality than state-of-art methods.
arXiv Detail & Related papers (2021-04-07T11:32:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.