Large Language and Text-to-3D Models for Engineering Design Optimization
- URL: http://arxiv.org/abs/2307.01230v1
- Date: Mon, 3 Jul 2023 07:54:09 GMT
- Title: Large Language and Text-to-3D Models for Engineering Design Optimization
- Authors: Thiago Rios, Stefan Menzel, Bernhard Sendhoff (Honda Research
Institute Europe)
- Abstract summary: We study the potential of deep text-to-3D models in the engineering domain.
We use Shap-E, a text-to-3D asset network by OpenAI, in the context of aerodynamic vehicle optimization.
- Score: 0.1740313383876245
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: The current advances in generative AI for learning large neural network
models with the capability to produce essays, images, music and even 3D assets
from text prompts create opportunities for a manifold of disciplines. In the
present paper, we study the potential of deep text-to-3D models in the
engineering domain, with focus on the chances and challenges when integrating
and interacting with 3D assets in computational simulation-based design
optimization. In contrast to traditional design optimization of 3D geometries
that often searches for the optimum designs using numerical representations,
such as B-Spline surface or deformation parameters in vehicle aerodynamic
optimization, natural language challenges the optimization framework by
requiring a different interpretation of variation operators while at the same
time may ease and motivate the human user interaction. Here, we propose and
realize a fully automated evolutionary design optimization framework using
Shap-E, a recently published text-to-3D asset network by OpenAI, in the context
of aerodynamic vehicle optimization. For representing text prompts in the
evolutionary optimization, we evaluate (a) a bag-of-words approach based on
prompt templates and Wordnet samples, and (b) a tokenisation approach based on
prompt templates and the byte pair encoding method from GPT4. Our main findings
from the optimizations indicate that, first, it is important to ensure that the
designs generated from prompts are within the object class of application, i.e.
diverse and novel designs need to be realistic, and, second, that more research
is required to develop methods where the strength of text prompt variations and
the resulting variations of the 3D designs share causal relations to some
degree to improve the optimization.
Related papers
- OrientDream: Streamlining Text-to-3D Generation with Explicit Orientation Control [66.03885917320189]
OrientDream is a camera orientation conditioned framework for efficient and multi-view consistent 3D generation from textual prompts.
Our strategy emphasizes the implementation of an explicit camera orientation conditioned feature in the pre-training of a 2D text-to-image diffusion module.
Our experiments reveal that our method not only produces high-quality NeRF models with consistent multi-view properties but also achieves an optimization speed significantly greater than existing methods.
arXiv Detail & Related papers (2024-06-14T13:16:18Z) - Generative AI-based Prompt Evolution Engineering Design Optimization With Vision-Language Model [22.535058343006828]
We present a prompt evolution design optimization (PEDO) framework contextualized in a vehicle design scenario.
We use a physics-based solver and a vision-language model for practical or functional guidance in the generated car designs.
Our investigations on a car design optimization problem show a wide spread of potential car designs generated at the early phase of the search.
arXiv Detail & Related papers (2024-06-13T14:11:19Z) - Text2VP: Generative AI for Visual Programming and Parametric Modeling [6.531561475204309]
This study creates and investigates an innovative application of generative AI in parametric modeling by leveraging a customized Text-to-Visual Programming (Text2VP) GPT derived from GPT-4.
The primary focus is on automating the generation of graph-based visual programming, including parameters and the links among the parameters, through AI-generated scripts.
Our testing demonstrates Text2VP's capability to generate working parametric models.
arXiv Detail & Related papers (2024-06-09T02:22:20Z) - Position: Leverage Foundational Models for Black-Box Optimization [19.583955195098497]
Large Language Models (LLMs) have stirred an extraordinary wave of innovation in the machine learning research domain.
We discuss the most promising ways foundational language models can revolutionize optimization.
arXiv Detail & Related papers (2024-05-06T15:10:46Z) - Multimodal Large Language Model is a Human-Aligned Annotator for Text-to-Image Generation [87.50120181861362]
VisionPrefer is a high-quality and fine-grained preference dataset that captures multiple preference aspects.
We train a reward model VP-Score over VisionPrefer to guide the training of text-to-image generative models and the preference prediction accuracy of VP-Score is comparable to human annotators.
arXiv Detail & Related papers (2024-04-23T14:53:15Z) - Compositional Generative Inverse Design [69.22782875567547]
Inverse design, where we seek to design input variables in order to optimize an underlying objective function, is an important problem.
We show that by instead optimizing over the learned energy function captured by the diffusion model, we can avoid such adversarial examples.
In an N-body interaction task and a challenging 2D multi-airfoil design task, we demonstrate that by composing the learned diffusion model at test time, our method allows us to design initial states and boundary shapes.
arXiv Detail & Related papers (2024-01-24T01:33:39Z) - Guide3D: Create 3D Avatars from Text and Image Guidance [55.71306021041785]
Guide3D is a text-and-image-guided generative model for 3D avatar generation based on diffusion models.
Our framework produces topologically and structurally correct geometry and high-resolution textures.
arXiv Detail & Related papers (2023-08-18T17:55:47Z) - ATT3D: Amortized Text-to-3D Object Synthesis [78.96673650638365]
We amortize optimization over text prompts by training on many prompts simultaneously with a unified model, instead of separately.
Our framework - Amortized text-to-3D (ATT3D) - enables knowledge-sharing between prompts to generalize to unseen setups and smooths between text for novel assets and simple animations.
arXiv Detail & Related papers (2023-06-06T17:59:10Z) - T2TD: Text-3D Generation Model based on Prior Knowledge Guidance [74.32278935880018]
We propose a novel text-3D generation model (T2TD), which introduces the related shapes or textual information as the prior knowledge to improve the performance of the 3D generation model.
Our approach significantly improves 3D model generation quality and outperforms the SOTA methods on the text2shape datasets.
arXiv Detail & Related papers (2023-05-25T06:05:52Z) - Early-Phase Performance-Driven Design using Generative Models [0.0]
This research introduces a novel method for performance-driven geometry generation that can afford interaction directly in the 3d modeling environment.
The method uses Machine Learning techniques to train a generative model offline.
By navigating the generative model's latent space, geometries with the desired characteristics can be quickly generated.
arXiv Detail & Related papers (2021-07-19T01:25:11Z) - Generative Design by Reinforcement Learning: Enhancing the Diversity of
Topology Optimization Designs [5.8010446129208155]
This study proposes a reinforcement learning based generative design process, with reward functions maximizing the diversity of topology designs.
We show that RL-based generative design produces a large number of diverse designs within a short inference time by exploiting GPU in a fully automated manner.
arXiv Detail & Related papers (2020-08-17T06:50:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.