Font Style that Fits an Image -- Font Generation Based on Image Context
- URL: http://arxiv.org/abs/2105.08879v1
- Date: Wed, 19 May 2021 01:53:04 GMT
- Title: Font Style that Fits an Image -- Font Generation Based on Image Context
- Authors: Taiga Miyazono, Brian Kenji Iwana, Daichi Haraguchi, Seiichi Uchida
- Abstract summary: We propose a method of generating a book title image based on its context within a book cover.
We propose an end-to-end neural network that inputs the book cover, a target location mask, and a desired book title and outputs stylized text suitable for the cover.
We demonstrate that the proposed method can effectively produce desirable and appropriate book cover text through quantitative and qualitative results.
- Score: 7.646713951724013
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: When fonts are used on documents, they are intentionally selected by
designers. For example, when designing a book cover, the typography of the text
is an important factor in the overall feel of the book. In addition, it needs
to be an appropriate font for the rest of the book cover. Thus, we propose a
method of generating a book title image based on its context within a book
cover. We propose an end-to-end neural network that inputs the book cover, a
target location mask, and a desired book title and outputs stylized text
suitable for the cover. The proposed network uses a combination of a
multi-input encoder-decoder, a text skeleton prediction network, a perception
network, and an adversarial discriminator. We demonstrate that the proposed
method can effectively produce desirable and appropriate book cover text
through quantitative and qualitative results.
Related papers
- Font Impression Estimation in the Wild [7.542892664684078]
We use a font dataset with annotation about font impressions and a convolutional neural network (CNN) framework for this task.
We propose an exemplar-based impression estimation approach, which relies on a strategy of ensembling impressions of exemplar fonts that are similar to the input image.
We conduct a correlation analysis between book genres and font impressions on real book cover images.
arXiv Detail & Related papers (2024-02-23T10:00:25Z) - VQ-Font: Few-Shot Font Generation with Structure-Aware Enhancement and
Quantization [52.870638830417]
We propose a VQGAN-based framework (i.e., VQ-Font) to enhance glyph fidelity through token prior refinement and structure-aware enhancement.
Specifically, we pre-train a VQGAN to encapsulate font token prior within a codebook. Subsequently, VQ-Font refines the synthesized glyphs with the codebook to eliminate the domain gap between synthesized and real-world strokes.
arXiv Detail & Related papers (2023-08-27T06:32:20Z) - Interleaving GANs with knowledge graphs to support design creativity for
book covers [77.34726150561087]
We apply Generative Adversarial Networks (GANs) to the book covers domain.
We interleave GANs with knowledge graphs to alter the input title to obtain multiple possible options for any given title.
Finally, we use the discriminator obtained during the training phase to select the best images generated with new titles.
arXiv Detail & Related papers (2023-08-03T08:56:56Z) - Unified Multi-Modal Latent Diffusion for Joint Subject and Text
Conditional Image Generation [63.061871048769596]
We present a novel Unified Multi-Modal Latent Diffusion (UMM-Diffusion) which takes joint texts and images containing specified subjects as input sequences.
To be more specific, both input texts and images are encoded into one unified multi-modal latent space.
Our method is able to generate high-quality images with complex semantics from both aspects of input texts and images.
arXiv Detail & Related papers (2023-03-16T13:50:20Z) - SpaText: Spatio-Textual Representation for Controllable Image Generation [61.89548017729586]
SpaText is a new method for text-to-image generation using open-vocabulary scene control.
In addition to a global text prompt that describes the entire scene, the user provides a segmentation map.
We show its effectiveness on two state-of-the-art diffusion models: pixel-based and latent-conditional-based.
arXiv Detail & Related papers (2022-11-25T18:59:10Z) - Book Cover Synthesis from the Summary [0.0]
We explore ways to produce a book cover using artificial intelligence based on the fact that there exists a relationship between the summary of the book and its cover.
We construct a dataset of English books that contains a large number of samples of summaries of existing books and their cover images.
We apply different text-to-image synthesis techniques to generate book covers from the summary and exhibit the results in this paper.
arXiv Detail & Related papers (2022-11-03T20:43:40Z) - A Multi-Implicit Neural Representation for Fonts [79.6123184198301]
font-specific discontinuities like edges and corners are difficult to represent using neural networks.
We introduce textitmulti-implicits to represent fonts as a permutation-in set of learned implict functions, without losing features.
arXiv Detail & Related papers (2021-06-12T21:40:11Z) - Towards Book Cover Design via Layout Graphs [18.028269880425455]
We propose a generative neural network that can produce book covers based on an easy-to-use layout graph.
The layout graph contains objects such as text, natural scene objects, and solid color spaces.
arXiv Detail & Related papers (2021-05-24T04:28:35Z) - Towards Open-World Text-Guided Face Image Generation and Manipulation [52.83401421019309]
We propose a unified framework for both face image generation and manipulation.
Our method supports open-world scenarios, including both image and text, without any re-training, fine-tuning, or post-processing.
arXiv Detail & Related papers (2021-04-18T16:56:07Z) - Deep multi-modal networks for book genre classification based on its
cover [0.0]
We propose a multi-modal deep learning framework to solve the cover-based book classification problem.
Our method adds an extra modality by extracting texts automatically from the book covers.
Results show that the multi-modal framework significantly outperforms the current state-of-the-art image-based models.
arXiv Detail & Related papers (2020-11-15T23:27:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.