Learning Aesthetic Layouts via Visual Guidance
- URL: http://arxiv.org/abs/2107.06262v1
- Date: Tue, 13 Jul 2021 17:46:42 GMT
- Title: Learning Aesthetic Layouts via Visual Guidance
- Authors: Qingyuan Zheng, Zhuoru Li, Adam Bargteil
- Abstract summary: We explore computational approaches for visual guidance to aid in creating pleasing art and graphic design.
We collected a dataset of art masterpieces and labeled the visual fixations with state-of-art vision models.
We clustered the visual guidance templates of the art masterpieces with unsupervised learning.
We show that the aesthetic visual guidance principles can be learned and integrated into a high-dimensional model and can be queried by the features of graphic elements.
- Score: 7.992550355579791
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: We explore computational approaches for visual guidance to aid in creating
aesthetically pleasing art and graphic design. Our work complements and builds
on previous work that developed models for how humans look at images. Our
approach comprises three steps. First, we collected a dataset of art
masterpieces and labeled the visual fixations with state-of-art vision models.
Second, we clustered the visual guidance templates of the art masterpieces with
unsupervised learning. Third, we developed a pipeline using generative
adversarial networks to learn the principles of visual guidance and that can
produce aesthetically pleasing layouts. We show that the aesthetic visual
guidance principles can be learned and integrated into a high-dimensional model
and can be queried by the features of graphic elements. We evaluate our
approach by generating layouts on various drawings and graphic designs.
Moreover, our model considers the color and structure of graphic elements when
generating layouts. Consequently, we believe our tool, which generates multiple
aesthetic layout options in seconds, can help artists create beautiful art and
graphic designs.
Related papers
- Automatic Layout Planning for Visually-Rich Documents with Instruction-Following Models [81.6240188672294]
In graphic design, non-professional users often struggle to create visually appealing layouts due to limited skills and resources.
We introduce a novel multimodal instruction-following framework for layout planning, allowing users to easily arrange visual elements into tailored layouts.
Our method not only simplifies the design process for non-professionals but also surpasses the performance of few-shot GPT-4V models, with mIoU higher by 12% on Crello.
arXiv Detail & Related papers (2024-04-23T17:58:33Z) - Impressions: Understanding Visual Semiotics and Aesthetic Impact [66.40617566253404]
We present Impressions, a novel dataset through which to investigate the semiotics of images.
We show that existing multimodal image captioning and conditional generation models struggle to simulate plausible human responses to images.
This dataset significantly improves their ability to model impressions and aesthetic evaluations of images through fine-tuning and few-shot adaptation.
arXiv Detail & Related papers (2023-10-27T04:30:18Z) - Composition-aware Graphic Layout GAN for Visual-textual Presentation
Designs [24.29890251913182]
We study the graphic layout generation problem of producing high-quality visual-textual presentation designs for given images.
We propose a deep generative model, dubbed as composition-aware graphic layout GAN (CGL-GAN), to synthesize layouts based on the global and spatial visual contents of input images.
arXiv Detail & Related papers (2022-04-30T16:42:13Z) - Infographics Wizard: Flexible Infographics Authoring and Design
Exploration [48.93421725740813]
Infographics are an aesthetic visual representation of information following specific design principles of human perception.
We propose a semi-automated infographic framework for general structured and flow-based infographic design generation.
For novice designers, our framework automatically creates and ranks infographic designs for a user-provided text with no requirement for design input.
We will also contribute an individual visual group (VG) designs dataset (in SVG) along with a 1k complete infographic image dataset with segmented VGs in this work.
arXiv Detail & Related papers (2022-04-21T06:26:06Z) - Detecting Visual Design Principles in Art and Architecture through Deep
Convolutional Neural Networks [0.0]
This research aims at a neural network model, which recognizes and classifies the design principles over different domains.
The proposed model learns from the knowledge of myriads of original designs, by capturing the underlying shared patterns.
arXiv Detail & Related papers (2021-08-09T14:00:17Z) - Learning Portrait Style Representations [34.59633886057044]
We study style representations learned by neural network architectures incorporating higher level characteristics.
We find variation in learned style features from incorporating triplets annotated by art historians as supervision for style similarity.
We also present the first large-scale dataset of portraits prepared for computational analysis.
arXiv Detail & Related papers (2020-12-08T01:36:45Z) - SketchEmbedNet: Learning Novel Concepts by Imitating Drawings [125.45799722437478]
We explore properties of image representations learned by training a model to produce sketches of images.
We show that this generative, class-agnostic model produces informative embeddings of images from novel examples, classes, and even novel datasets in a few-shot setting.
arXiv Detail & Related papers (2020-08-27T16:43:28Z) - Modeling Artistic Workflows for Image Generation and Editing [83.43047077223947]
We propose a generative model that follows a given artistic workflow.
It enables both multi-stage image generation as well as multi-stage image editing of an existing piece of art.
arXiv Detail & Related papers (2020-07-14T17:54:26Z) - Deep Plastic Surgery: Robust and Controllable Image Editing with
Human-Drawn Sketches [133.01690754567252]
Sketch-based image editing aims to synthesize and modify photos based on the structural information provided by the human-drawn sketches.
Deep Plastic Surgery is a novel, robust and controllable image editing framework that allows users to interactively edit images using hand-drawn sketch inputs.
arXiv Detail & Related papers (2020-01-09T08:57:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.