Book Cover Synthesis from the Summary
- URL: http://arxiv.org/abs/2211.02138v1
- Date: Thu, 3 Nov 2022 20:43:40 GMT
- Title: Book Cover Synthesis from the Summary
- Authors: Emdadul Haque, Md. Faraz Kabir Khan, Mohammad Imrul Jubair, Jarin
Anjum, Abrar Zahir Niloy
- Abstract summary: We explore ways to produce a book cover using artificial intelligence based on the fact that there exists a relationship between the summary of the book and its cover.
We construct a dataset of English books that contains a large number of samples of summaries of existing books and their cover images.
We apply different text-to-image synthesis techniques to generate book covers from the summary and exhibit the results in this paper.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The cover is the face of a book and is a point of attraction for the readers.
Designing book covers is an essential task in the publishing industry. One of
the main challenges in creating a book cover is representing the theme of the
book's content in a single image. In this research, we explore ways to produce
a book cover using artificial intelligence based on the fact that there exists
a relationship between the summary of the book and its cover. Our key
motivation is the application of text-to-image synthesis methods to generate
images from given text or captions. We explore several existing text-to-image
conversion techniques for this purpose and propose an approach to exploit these
frameworks for producing book covers from provided summaries. We construct a
dataset of English books that contains a large number of samples of summaries
of existing books and their cover images. In this paper, we describe our
approach to collecting, organizing, and pre-processing the dataset to use it
for training models. We apply different text-to-image synthesis techniques to
generate book covers from the summary and exhibit the results in this paper.
Related papers
- Decoupled Textual Embeddings for Customized Image Generation [62.98933630971543]
Customized text-to-image generation aims to learn user-specified concepts with a few images.
Existing methods usually suffer from overfitting issues and entangle the subject-unrelated information with the learned concept.
We propose the DETEX, a novel approach that learns the disentangled concept embedding for flexible customized text-to-image generation.
arXiv Detail & Related papers (2023-12-19T03:32:10Z) - Interleaving GANs with knowledge graphs to support design creativity for
book covers [77.34726150561087]
We apply Generative Adversarial Networks (GANs) to the book covers domain.
We interleave GANs with knowledge graphs to alter the input title to obtain multiple possible options for any given title.
Finally, we use the discriminator obtained during the training phase to select the best images generated with new titles.
arXiv Detail & Related papers (2023-08-03T08:56:56Z) - Break-A-Scene: Extracting Multiple Concepts from a Single Image [80.47666266017207]
We introduce the task of textual scene decomposition.
We propose augmenting the input image with masks that indicate the presence of target concepts.
We then present a novel two-phase customization process.
arXiv Detail & Related papers (2023-05-25T17:59:04Z) - Enhancing Textbooks with Visuals from the Web for Improved Learning [50.01434477801967]
In this paper, we investigate the effectiveness of vision-language models to automatically enhance textbooks with images from the web.
We collect a dataset of e-textbooks in the math, science, social science and business domains.
We then set up a text-image matching task that involves retrieving and appropriately assigning web images to textbooks.
arXiv Detail & Related papers (2023-04-18T12:16:39Z) - Text-to-image Diffusion Models in Generative AI: A Survey [75.32882187215394]
We present a review of state-of-the-art methods on text-conditioned image synthesis, i.e., text-to-image.
We discuss applications beyond text-to-image generation: text-guided creative generation and text-guided image editing.
arXiv Detail & Related papers (2023-03-14T13:49:54Z) - CREATIVESUMM: Shared Task on Automatic Summarization for Creative
Writing [90.58269243992318]
This paper introduces the shared task of summarizing documents in several creative domains, namely literary texts, movie scripts, and television scripts.
We introduce four sub-tasks and their corresponding datasets, focusing on summarizing books, movie scripts, primetime television scripts, and daytime soap opera scripts.
As part of the CREATIVESUMM workshop at COLING 2022, the shared task attracted 18 submissions in total.
arXiv Detail & Related papers (2022-11-10T21:31:03Z) - Towards Book Cover Design via Layout Graphs [18.028269880425455]
We propose a generative neural network that can produce book covers based on an easy-to-use layout graph.
The layout graph contains objects such as text, natural scene objects, and solid color spaces.
arXiv Detail & Related papers (2021-05-24T04:28:35Z) - Font Style that Fits an Image -- Font Generation Based on Image Context [7.646713951724013]
We propose a method of generating a book title image based on its context within a book cover.
We propose an end-to-end neural network that inputs the book cover, a target location mask, and a desired book title and outputs stylized text suitable for the cover.
We demonstrate that the proposed method can effectively produce desirable and appropriate book cover text through quantitative and qualitative results.
arXiv Detail & Related papers (2021-05-19T01:53:04Z) - Deep multi-modal networks for book genre classification based on its
cover [0.0]
We propose a multi-modal deep learning framework to solve the cover-based book classification problem.
Our method adds an extra modality by extracting texts automatically from the book covers.
Results show that the multi-modal framework significantly outperforms the current state-of-the-art image-based models.
arXiv Detail & Related papers (2020-11-15T23:27:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.