Learning Perceptual Manifold of Fonts
- URL: http://arxiv.org/abs/2106.09198v1
- Date: Thu, 17 Jun 2021 01:22:52 GMT
- Title: Learning Perceptual Manifold of Fonts
- Authors: Haoran Xie and Yuki Fujita and Kazunori Miyata
- Abstract summary: We propose the perceptual manifold of fonts to visualize the perceptual adjustment in the latent space of a generative model of fonts.
In contrast to the conventional user interface in our study, the proposed font-exploring user interface is efficient and helpful in the designated user preference.
- Score: 7.395615703126767
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Along the rapid development of deep learning techniques in generative models,
it is becoming an urgent issue to combine machine intelligence with human
intelligence to solve the practical applications. Motivated by this
methodology, this work aims to adjust the machine generated character fonts
with the effort of human workers in the perception study. Although numerous
fonts are available online for public usage, it is difficult and challenging to
generate and explore a font to meet the preferences for common users. To solve
the specific issue, we propose the perceptual manifold of fonts to visualize
the perceptual adjustment in the latent space of a generative model of fonts.
In our framework, we adopt the variational autoencoder network for the font
generation. Then, we conduct a perceptual study on the generated fonts from the
multi-dimensional latent space of the generative model. After we obtained the
distribution data of specific preferences, we utilize manifold learning
approach to visualize the font distribution. In contrast to the conventional
user interface in our user study, the proposed font-exploring user interface is
efficient and helpful in the designated user preference.
Related papers
- GRIF-DM: Generation of Rich Impression Fonts using Diffusion Models [18.15911470339845]
We introduce a diffusion-based method, termed ourmethod, to generate fonts that vividly embody specific impressions.
Our experimental results, conducted on the MyFonts dataset, affirm that this method is capable of producing realistic, vibrant, and high-fidelity fonts.
arXiv Detail & Related papers (2024-08-14T02:26:46Z) - JeDi: Joint-Image Diffusion Models for Finetuning-Free Personalized Text-to-Image Generation [49.997839600988875]
Existing personalization methods rely on finetuning a text-to-image foundation model on a user's custom dataset.
We propose Joint-Image Diffusion (jedi), an effective technique for learning a finetuning-free personalization model.
Our model achieves state-of-the-art generation quality, both quantitatively and qualitatively, significantly outperforming both the prior finetuning-based and finetuning-free personalization baselines.
arXiv Detail & Related papers (2024-07-08T17:59:02Z) - FontStudio: Shape-Adaptive Diffusion Model for Coherent and Consistent Font Effect Generation [38.730628018627975]
This research aims to tackle the generation of text effects for multilingual fonts.
We introduce a novel shape-adaptive diffusion model capable of interpreting the given shape.
We also present a training-free, shape-adaptive effect transfer method for transferring textures from a generated reference letter to others.
arXiv Detail & Related papers (2024-06-12T16:43:47Z) - Multimodal Large Language Model is a Human-Aligned Annotator for Text-to-Image Generation [87.50120181861362]
VisionPrefer is a high-quality and fine-grained preference dataset that captures multiple preference aspects.
We train a reward model VP-Score over VisionPrefer to guide the training of text-to-image generative models and the preference prediction accuracy of VP-Score is comparable to human annotators.
arXiv Detail & Related papers (2024-04-23T14:53:15Z) - Diff-Font: Diffusion Model for Robust One-Shot Font Generation [110.45944936952309]
We propose a novel one-shot font generation method based on a diffusion model, named Diff-Font.
The proposed model aims to generate the entire font library by giving only one sample as the reference.
The well-trained Diff-Font is not only robust to font gap and font variation, but also achieved promising performance on difficult character generation.
arXiv Detail & Related papers (2022-12-12T13:51:50Z) - FontNet: Closing the gap to font designer performance in font synthesis [3.991334489146843]
We propose a model, called FontNet, that learns to separate font styles in the embedding space where distances directly correspond to a measure of font similarity.
We design the network architecture and training procedure that can be adopted for any language system and can produce high-resolution font images.
arXiv Detail & Related papers (2022-05-13T08:37:10Z) - Scalable Font Reconstruction with Dual Latent Manifolds [55.29525824849242]
We propose a deep generative model that performs typography analysis and font reconstruction.
Our approach enables us to massively scale up the number of character types we can effectively model.
We evaluate on the task of font reconstruction over various datasets representing character types of many languages.
arXiv Detail & Related papers (2021-09-10T20:37:43Z) - A Multi-Implicit Neural Representation for Fonts [79.6123184198301]
font-specific discontinuities like edges and corners are difficult to represent using neural networks.
We introduce textitmulti-implicits to represent fonts as a permutation-in set of learned implict functions, without losing features.
arXiv Detail & Related papers (2021-06-12T21:40:11Z) - AdaptiFont: Increasing Individuals' Reading Speed with a Generative Font
Model and Bayesian Optimization [3.480626767752489]
AdaptiFont is a human-in-the-loop system aimed at interactively increasing readability of text displayed on a monitor.
We generate new true-type-fonts through active learning, render texts with the new font, and measure individual users' reading speed.
The results of a user study show that this adaptive font generation system finds regions in the font space corresponding to high reading speeds, that these fonts significantly increase participants' reading speed, and that the found fonts are significantly different across individual readers.
arXiv Detail & Related papers (2021-04-21T19:56:28Z) - Let Me Choose: From Verbal Context to Font Selection [50.293897197235296]
We aim to learn associations between visual attributes of fonts and the verbal context of the texts they are typically applied to.
We introduce a new dataset, containing examples of different topics in social media posts and ads, labeled through crowd-sourcing.
arXiv Detail & Related papers (2020-05-03T17:36:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.