Few-Shot Font Generation by Learning Fine-Grained Local Styles
- URL: http://arxiv.org/abs/2205.09965v2
- Date: Mon, 23 May 2022 13:20:31 GMT
- Title: Few-Shot Font Generation by Learning Fine-Grained Local Styles
- Authors: Licheng Tang, Yiyang Cai, Jiaming Liu, Zhibin Hong, Mingming Gong,
Minhu Fan, Junyu Han, Jingtuo Liu, Errui Ding, Jingdong Wang
- Abstract summary: Few-shot font generation (FFG) aims to generate a new font with a few examples.
We propose a new font generation approach by learning 1) the fine-grained local styles from references, and 2) the spatial correspondence between the content and reference glyphs.
- Score: 90.39288370855115
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Few-shot font generation (FFG), which aims to generate a new font with a few
examples, is gaining increasing attention due to the significant reduction in
labor cost. A typical FFG pipeline considers characters in a standard font
library as content glyphs and transfers them to a new target font by extracting
style information from the reference glyphs. Most existing solutions explicitly
disentangle content and style of reference glyphs globally or component-wisely.
However, the style of glyphs mainly lies in the local details, i.e. the styles
of radicals, components, and strokes together depict the style of a glyph.
Therefore, even a single character can contain different styles distributed
over spatial locations. In this paper, we propose a new font generation
approach by learning 1) the fine-grained local styles from references, and 2)
the spatial correspondence between the content and reference glyphs. Therefore,
each spatial location in the content glyph can be assigned with the right
fine-grained style. To this end, we adopt cross-attention over the
representation of the content glyphs as the queries and the representations of
the reference glyphs as the keys and values. Instead of explicitly
disentangling global or component-wise modeling, the cross-attention mechanism
can attend to the right local styles in the reference glyphs and aggregate the
reference styles into a fine-grained style representation for the given content
glyphs. The experiments show that the proposed method outperforms the
state-of-the-art methods in FFG. In particular, the user studies also
demonstrate the style consistency of our approach significantly outperforms
previous methods.
Related papers
- Few shot font generation via transferring similarity guided global style
and quantization local style [11.817299400850176]
We present a novel font generation approach by aggregating styles from character similarity-guided global features and stylized component-level representations.
Our AFFG method could obtain a complete set of component-level style representations, and also control the global glyph characteristics.
arXiv Detail & Related papers (2023-09-02T05:05:40Z) - VQ-Font: Few-Shot Font Generation with Structure-Aware Enhancement and
Quantization [52.870638830417]
We propose a VQGAN-based framework (i.e., VQ-Font) to enhance glyph fidelity through token prior refinement and structure-aware enhancement.
Specifically, we pre-train a VQGAN to encapsulate font token prior within a codebook. Subsequently, VQ-Font refines the synthesized glyphs with the codebook to eliminate the domain gap between synthesized and real-world strokes.
arXiv Detail & Related papers (2023-08-27T06:32:20Z) - CF-Font: Content Fusion for Few-shot Font Generation [63.79915037830131]
We propose a content fusion module (CFM) to project the content feature into a linear space defined by the content features of basis fonts.
Our method also allows to optimize the style representation vector of reference images.
We have evaluated our method on a dataset of 300 fonts with 6.5k characters each.
arXiv Detail & Related papers (2023-03-24T14:18:40Z) - Few-shot Font Generation by Learning Style Difference and Similarity [84.76381937516356]
We propose a novel font generation approach by learning the Difference between different styles and the Similarity of the same style (DS-Font)
Specifically, we propose a multi-layer style projector for style encoding and realize a distinctive style representation via our proposed Cluster-level Contrastive Style (CCS) loss.
arXiv Detail & Related papers (2023-01-24T13:57:25Z) - XMP-Font: Self-Supervised Cross-Modality Pre-training for Few-Shot Font
Generation [13.569449355929574]
We propose a self-supervised cross-modality pre-training strategy and a cross-modality transformer-based encoder.
The encoder is conditioned jointly on the glyph image and the corresponding stroke labels.
It only requires one reference glyph and achieves the lowest rate of bad cases in the few-shot font generation task 28% lower than the second best.
arXiv Detail & Related papers (2022-04-11T13:34:40Z) - Few-shot Font Generation with Weakly Supervised Localized
Representations [17.97183447033118]
We propose a novel font generation method that learns localized styles, namely component-wise style representations, instead of universal styles.
Our method shows remarkably better few-shot font generation results (with only eight reference glyphs) than other state-of-the-art methods.
arXiv Detail & Related papers (2021-12-22T14:26:53Z) - Font Completion and Manipulation by Cycling Between Multi-Modality
Representations [113.26243126754704]
We innovate to explore the generation of font glyphs as 2D graphic objects with the graph as an intermediate representation.
We formulate a cross-modality cycled image-to-image structure with a graph between an image encoder and an image.
Our model generates improved results than both image-to-image baseline and previous state-of-the-art methods for glyph completion.
arXiv Detail & Related papers (2021-08-30T02:43:29Z) - A Multi-Implicit Neural Representation for Fonts [79.6123184198301]
font-specific discontinuities like edges and corners are difficult to represent using neural networks.
We introduce textitmulti-implicits to represent fonts as a permutation-in set of learned implict functions, without losing features.
arXiv Detail & Related papers (2021-06-12T21:40:11Z) - Few-shot Font Generation with Localized Style Representations and
Factorization [23.781619323447003]
We propose a novel font generation method by learning localized styles, namely component-wise style representations, instead of universal styles.
Our method shows remarkably better few-shot font generation results (with only 8 reference glyph images) than other state-of-the-arts.
arXiv Detail & Related papers (2020-09-23T10:33:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.