DeepVecFont-v2: Exploiting Transformers to Synthesize Vector Fonts with
Higher Quality
- URL: http://arxiv.org/abs/2303.14585v1
- Date: Sat, 25 Mar 2023 23:28:19 GMT
- Title: DeepVecFont-v2: Exploiting Transformers to Synthesize Vector Fonts with
Higher Quality
- Authors: Yuqing Wang, Yizhi Wang, Longhui Yu, Yuesheng Zhu, Zhouhui Lian
- Abstract summary: This paper proposes an enhanced version of DeepVecFont for vector font synthesis.
We adopt Transformers instead of RNNs to process sequential data and design a relaxation representation for vector outlines.
We also propose to sample auxiliary points in addition to control points to precisely align the generated and target B'ezier curves or lines.
- Score: 38.32966391626858
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Vector font synthesis is a challenging and ongoing problem in the fields of
Computer Vision and Computer Graphics. The recently-proposed DeepVecFont
achieved state-of-the-art performance by exploiting information of both the
image and sequence modalities of vector fonts. However, it has limited
capability for handling long sequence data and heavily relies on an
image-guided outline refinement post-processing. Thus, vector glyphs
synthesized by DeepVecFont still often contain some distortions and artifacts
and cannot rival human-designed results. To address the above problems, this
paper proposes an enhanced version of DeepVecFont mainly by making the
following three novel technical contributions. First, we adopt Transformers
instead of RNNs to process sequential data and design a relaxation
representation for vector outlines, markedly improving the model's capability
and stability of synthesizing long and complex outlines. Second, we propose to
sample auxiliary points in addition to control points to precisely align the
generated and target B\'ezier curves or lines. Finally, to alleviate error
accumulation in the sequential generation process, we develop a context-based
self-refinement module based on another Transformer-based decoder to remove
artifacts in the initially synthesized glyphs. Both qualitative and
quantitative results demonstrate that the proposed method effectively resolves
those intrinsic problems of the original DeepVecFont and outperforms existing
approaches in generating English and Chinese vector fonts with complicated
structures and diverse styles.
Related papers
- HFH-Font: Few-shot Chinese Font Synthesis with Higher Quality, Faster Speed, and Higher Resolution [17.977410216055024]
We introduce HFH-Font, a few-shot font synthesis method capable of efficiently generating high-resolution glyph images.
For the first time, large-scale Chinese vector fonts of a quality comparable to those manually created by professional font designers can be automatically generated.
arXiv Detail & Related papers (2024-10-09T02:30:24Z) - SuperSVG: Superpixel-based Scalable Vector Graphics Synthesis [66.44553285020066]
SuperSVG is a superpixel-based vectorization model that achieves fast and high-precision image vectorization.
We propose a two-stage self-training framework, where a coarse-stage model is employed to reconstruct the main structure and a refinement-stage model is used for enriching the details.
Experiments demonstrate the superior performance of our method in terms of reconstruction accuracy and inference time compared to state-of-the-art approaches.
arXiv Detail & Related papers (2024-06-14T07:43:23Z) - Texture, Shape and Order Matter: A New Transformer Design for Sequential DeepFake Detection [57.100891917805086]
Sequential DeepFake detection is an emerging task that predicts the manipulation sequence in order.
This paper describes a new Transformer design, called TSOM, by exploring three perspectives: Texture, Shape, and Order of Manipulations.
arXiv Detail & Related papers (2024-04-22T04:47:52Z) - VQ-Font: Few-Shot Font Generation with Structure-Aware Enhancement and
Quantization [52.870638830417]
We propose a VQGAN-based framework (i.e., VQ-Font) to enhance glyph fidelity through token prior refinement and structure-aware enhancement.
Specifically, we pre-train a VQGAN to encapsulate font token prior within a codebook. Subsequently, VQ-Font refines the synthesized glyphs with the codebook to eliminate the domain gap between synthesized and real-world strokes.
arXiv Detail & Related papers (2023-08-27T06:32:20Z) - DualVector: Unsupervised Vector Font Synthesis with Dual-Part
Representation [43.64428946288288]
Current font synthesis methods fail to represent the shape concisely or require vector supervision during training.
We propose a novel dual-part representation for vector glyphs, where each glyph is modeled as a collection of closed "positive" and "negative" path pairs.
Our method, named Dual-of-Font-art, outperforms state-of-the-art methods for practical use.
arXiv Detail & Related papers (2023-05-17T08:18:06Z) - VecFontSDF: Learning to Reconstruct and Synthesize High-quality Vector
Fonts via Signed Distance Functions [15.47282857047361]
This paper proposes an end-to-end trainable method, VecFontSDF, to reconstruct and synthesize high-quality vector fonts.
Based on the proposed SDF-based implicit shape representation, VecFontSDF learns to model each glyph as shape primitives enclosed by several parabolic curves.
arXiv Detail & Related papers (2023-03-22T16:14:39Z) - Diff-Font: Diffusion Model for Robust One-Shot Font Generation [110.45944936952309]
We propose a novel one-shot font generation method based on a diffusion model, named Diff-Font.
The proposed model aims to generate the entire font library by giving only one sample as the reference.
The well-trained Diff-Font is not only robust to font gap and font variation, but also achieved promising performance on difficult character generation.
arXiv Detail & Related papers (2022-12-12T13:51:50Z) - FontTransformer: Few-shot High-resolution Chinese Glyph Image Synthesis
via Stacked Transformers [21.705680113996742]
This paper proposes FontTransformer, a novel few-shot learning model, for high-resolution Chinese glyph image synthesis.
We also design a novel encoding scheme to feed more glyph information and prior knowledge to our model.
arXiv Detail & Related papers (2022-10-12T15:09:22Z) - DeepVecFont: Synthesizing High-quality Vector Fonts via Dual-modality
Learning [21.123297001902177]
We propose a novel method, DeepVecFont, to generate visually-pleasing vector glyphs.
The highlights of this paper are threefold. First, we design a dual-modality learning strategy which utilizes both image-aspect and sequence-aspect features of fonts to synthesize vector glyphs.
Second, we provide a new generative paradigm to handle unstructured data (e.g., vector glyphs) by randomly sampling plausible results to get the optimal one which is further refined under the guidance of generated structured data.
arXiv Detail & Related papers (2021-10-13T12:57:19Z) - Text Compression-aided Transformer Encoding [77.16960983003271]
We propose explicit and implicit text compression approaches to enhance the Transformer encoding.
backbone information, meaning the gist of the input text, is not specifically focused on.
Our evaluation on benchmark datasets shows that the proposed explicit and implicit text compression approaches improve results in comparison to strong baselines.
arXiv Detail & Related papers (2021-02-11T11:28:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.