ChatPattern: Layout Pattern Customization via Natural Language
- URL: http://arxiv.org/abs/2403.15434v1
- Date: Fri, 15 Mar 2024 09:15:22 GMT
- Title: ChatPattern: Layout Pattern Customization via Natural Language
- Authors: Zixiao Wang, Yunheng Shen, Xufeng Yao, Wenqian Zhao, Yang Bai, Farzan Farnia, Bei Yu,
- Abstract summary: ChatPattern is a novel Large-Language-Model powered framework for flexible pattern customization.
LLM agent can interpret natural language requirements and operate design tools to meet specified needs.
generator excels in conditional layout generation, pattern modification, and memory-friendly patterns extension.
- Score: 18.611898021267923
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Existing works focus on fixed-size layout pattern generation, while the more practical free-size pattern generation receives limited attention. In this paper, we propose ChatPattern, a novel Large-Language-Model (LLM) powered framework for flexible pattern customization. ChatPattern utilizes a two-part system featuring an expert LLM agent and a highly controllable layout pattern generator. The LLM agent can interpret natural language requirements and operate design tools to meet specified needs, while the generator excels in conditional layout generation, pattern modification, and memory-friendly patterns extension. Experiments on challenging pattern generation setting shows the ability of ChatPattern to synthesize high-quality large-scale patterns.
Related papers
- PatternPaint: Generating Layout Patterns Using Generative AI and Inpainting Techniques [5.126358554705107]
Existing training-based ML pattern generation approaches struggle to produce legal layout patterns in the early stages of technology node development.
We propose PatternPaint, a training-free framework capable of generating legal patterns with limited DRC Clean training samples.
PatternPaint is the first framework to generate a complex 2D layout pattern library using only 20 design rule clean layout patterns as input.
arXiv Detail & Related papers (2024-09-02T16:02:26Z) - PosterLLaVa: Constructing a Unified Multi-modal Layout Generator with LLM [58.67882997399021]
Our research introduces a unified framework for automated graphic layout generation.
Our data-driven method employs structured text (JSON format) and visual instruction tuning to generate layouts.
We conduct extensive experiments and achieved state-of-the-art (SOTA) performance on public multi-modal layout generation benchmarks.
arXiv Detail & Related papers (2024-06-05T03:05:52Z) - TextDiffuser-2: Unleashing the Power of Language Models for Text
Rendering [118.30923824681642]
TextDiffuser-2 aims to unleash the power of language models for text rendering.
We utilize the language model within the diffusion model to encode the position and texts at the line level.
We conduct extensive experiments and incorporate user studies involving human participants as well as GPT-4V.
arXiv Detail & Related papers (2023-11-28T04:02:40Z) - FiLM: Fill-in Language Models for Any-Order Generation [71.42044325886194]
Fill-in Language Model (FiLM) is a new language modeling approach that allows for flexible generation at any position without adhering to a specific generation order.
During inference, FiLM can seamlessly insert missing phrases, sentences, or paragraphs.
FiLM outperforms existing infilling methods that rely on left-to-right language models trained on rearranged text segments.
arXiv Detail & Related papers (2023-10-15T19:37:39Z) - PatternGPT :A Pattern-Driven Framework for Large Language Model Text
Generation [1.7259824817932292]
This paper proposes PatternGPT, a pattern-driven text generation framework for Large Language Models.
The framework utilizes the extraction capability of Large Language Models to generate rich and diversified structured and formalized patterns.
external knowledge such as judgment criteria and optimization algorithms are used to search for high-quality patterns.
arXiv Detail & Related papers (2023-07-02T04:32:41Z) - Generating Images with Multimodal Language Models [78.6660334861137]
We propose a method to fuse frozen text-only large language models with pre-trained image encoder and decoder models.
Our model demonstrates a wide suite of multimodal capabilities: image retrieval, novel image generation, and multimodal dialogue.
arXiv Detail & Related papers (2023-05-26T19:22:03Z) - LayoutGPT: Compositional Visual Planning and Generation with Large
Language Models [98.81962282674151]
Large Language Models (LLMs) can serve as visual planners by generating layouts from text conditions.
We propose LayoutGPT, a method to compose in-context visual demonstrations in style sheet language.
arXiv Detail & Related papers (2023-05-24T17:56:16Z) - Text-Blueprint: An Interactive Platform for Plan-based Conditional
Generation [84.95981645040281]
Planning can be a useful intermediate step to render conditional generation less opaque and more grounded.
We present a web browser-based demonstration for query-focused summarization that uses a sequence of question-answer pairs.
arXiv Detail & Related papers (2023-04-28T18:14:48Z) - DiffPattern: Layout Pattern Generation via Discrete Diffusion [16.148506119712735]
We propose toolDiffPattern to generate reliable layout patterns.
Our experiments on several benchmark settings show that toolDiffPattern significantly outperforms existing baselines.
arXiv Detail & Related papers (2023-03-23T06:16:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.