Learning Aesthetic Layouts via Visual Guidance
- URL: http://arxiv.org/abs/2107.06262v1
- Date: Tue, 13 Jul 2021 17:46:42 GMT
- Title: Learning Aesthetic Layouts via Visual Guidance
- Authors: Qingyuan Zheng, Zhuoru Li, Adam Bargteil
- Abstract summary: We explore computational approaches for visual guidance to aid in creating pleasing art and graphic design.
We collected a dataset of art masterpieces and labeled the visual fixations with state-of-art vision models.
We clustered the visual guidance templates of the art masterpieces with unsupervised learning.
We show that the aesthetic visual guidance principles can be learned and integrated into a high-dimensional model and can be queried by the features of graphic elements.
- Score: 7.992550355579791
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: We explore computational approaches for visual guidance to aid in creating
aesthetically pleasing art and graphic design. Our work complements and builds
on previous work that developed models for how humans look at images. Our
approach comprises three steps. First, we collected a dataset of art
masterpieces and labeled the visual fixations with state-of-art vision models.
Second, we clustered the visual guidance templates of the art masterpieces with
unsupervised learning. Third, we developed a pipeline using generative
adversarial networks to learn the principles of visual guidance and that can
produce aesthetically pleasing layouts. We show that the aesthetic visual
guidance principles can be learned and integrated into a high-dimensional model
and can be queried by the features of graphic elements. We evaluate our
approach by generating layouts on various drawings and graphic designs.
Moreover, our model considers the color and structure of graphic elements when
generating layouts. Consequently, we believe our tool, which generates multiple
aesthetic layout options in seconds, can help artists create beautiful art and
graphic designs.
Related papers
- Diffusion-Based Visual Art Creation: A Survey and New Perspectives [51.522935314070416]
This survey explores the emerging realm of diffusion-based visual art creation, examining its development from both artistic and technical perspectives.
Our findings reveal how artistic requirements are transformed into technical challenges and highlight the design and application of diffusion-based methods within visual art creation.
We aim to shed light on the mechanisms through which AI systems emulate and possibly, enhance human capacities in artistic perception and creativity.
arXiv Detail & Related papers (2024-08-22T04:49:50Z) - GalleryGPT: Analyzing Paintings with Large Multimodal Models [64.98398357569765]
Artwork analysis is important and fundamental skill for art appreciation, which could enrich personal aesthetic sensibility and facilitate the critical thinking ability.
Previous works for automatically analyzing artworks mainly focus on classification, retrieval, and other simple tasks, which is far from the goal of AI.
We introduce a superior large multimodal model for painting analysis composing, dubbed GalleryGPT, which is slightly modified and fine-tuned based on LLaVA architecture.
arXiv Detail & Related papers (2024-08-01T11:52:56Z) - BlenderAlchemy: Editing 3D Graphics with Vision-Language Models [4.852796482609347]
A vision-based edit generator and state evaluator work together to find the correct sequence of actions to achieve the goal.
Inspired by the role of visual imagination in the human design process, we supplement the visual reasoning capabilities of Vision-Language Models with "imagined" reference images.
arXiv Detail & Related papers (2024-04-26T19:37:13Z) - Automatic Layout Planning for Visually-Rich Documents with Instruction-Following Models [81.6240188672294]
In graphic design, non-professional users often struggle to create visually appealing layouts due to limited skills and resources.
We introduce a novel multimodal instruction-following framework for layout planning, allowing users to easily arrange visual elements into tailored layouts.
Our method not only simplifies the design process for non-professionals but also surpasses the performance of few-shot GPT-4V models, with mIoU higher by 12% on Crello.
arXiv Detail & Related papers (2024-04-23T17:58:33Z) - Impressions: Understanding Visual Semiotics and Aesthetic Impact [66.40617566253404]
We present Impressions, a novel dataset through which to investigate the semiotics of images.
We show that existing multimodal image captioning and conditional generation models struggle to simulate plausible human responses to images.
This dataset significantly improves their ability to model impressions and aesthetic evaluations of images through fine-tuning and few-shot adaptation.
arXiv Detail & Related papers (2023-10-27T04:30:18Z) - Composition-aware Graphic Layout GAN for Visual-textual Presentation
Designs [24.29890251913182]
We study the graphic layout generation problem of producing high-quality visual-textual presentation designs for given images.
We propose a deep generative model, dubbed as composition-aware graphic layout GAN (CGL-GAN), to synthesize layouts based on the global and spatial visual contents of input images.
arXiv Detail & Related papers (2022-04-30T16:42:13Z) - Detecting Visual Design Principles in Art and Architecture through Deep
Convolutional Neural Networks [0.0]
This research aims at a neural network model, which recognizes and classifies the design principles over different domains.
The proposed model learns from the knowledge of myriads of original designs, by capturing the underlying shared patterns.
arXiv Detail & Related papers (2021-08-09T14:00:17Z) - SketchEmbedNet: Learning Novel Concepts by Imitating Drawings [125.45799722437478]
We explore properties of image representations learned by training a model to produce sketches of images.
We show that this generative, class-agnostic model produces informative embeddings of images from novel examples, classes, and even novel datasets in a few-shot setting.
arXiv Detail & Related papers (2020-08-27T16:43:28Z) - Deep Plastic Surgery: Robust and Controllable Image Editing with
Human-Drawn Sketches [133.01690754567252]
Sketch-based image editing aims to synthesize and modify photos based on the structural information provided by the human-drawn sketches.
Deep Plastic Surgery is a novel, robust and controllable image editing framework that allows users to interactively edit images using hand-drawn sketch inputs.
arXiv Detail & Related papers (2020-01-09T08:57:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.