Three-dimensional Deep Shape Optimization with a Limited Dataset
- URL: http://arxiv.org/abs/2506.12326v1
- Date: Sat, 14 Jun 2025 03:26:28 GMT
- Title: Three-dimensional Deep Shape Optimization with a Limited Dataset
- Authors: Yongmin Kwon, Namwoo Kang,
- Abstract summary: Generative models have attracted considerable attention for their ability to produce novel shapes.<n>Their application in mechanical design remains constrained due to the limited size and variability of available datasets.<n>This study proposes a deep learning-based optimization framework specifically tailored for shape optimization with limited datasets.
- Score: 1.6574413179773761
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Generative models have attracted considerable attention for their ability to produce novel shapes. However, their application in mechanical design remains constrained due to the limited size and variability of available datasets. This study proposes a deep learning-based optimization framework specifically tailored for shape optimization with limited datasets, leveraging positional encoding and a Lipschitz regularization term to robustly learn geometric characteristics and maintain a meaningful latent space. Through extensive experiments, the proposed approach demonstrates robustness, generalizability and effectiveness in addressing typical limitations of conventional optimization frameworks. The validity of the methodology is confirmed through multi-objective shape optimization experiments conducted on diverse three-dimensional datasets, including wheels and cars, highlighting the model's versatility in producing practical and high-quality design outcomes even under data-constrained conditions.
Related papers
- Efficient Design of Compliant Mechanisms Using Multi-Objective Optimization [50.24983453990065]
We address the synthesis of a compliant cross-hinge mechanism capable of large angular strokes.<n>We formulate a multi-objective optimization problem based on kinetostatic performance measures.
arXiv Detail & Related papers (2025-04-23T06:29:10Z) - Generative Topology Optimization: Exploring Diverse Solutions in Structural Design [11.585099298173285]
Topology optimization (TO) is a family of computational methods that derive near-optimal geometries from formal problem descriptions.<n>We introduce Generative Topology Optimization (GenTO) - a data-free method that trains a neural network to generate structurally compliant shapes.<n>We show that GenTO produces more diverse solutions than any prior method while maintaining near-optimality.
arXiv Detail & Related papers (2025-02-17T21:24:18Z) - Implicitly Guided Design with PropEn: Match your Data to Follow the Gradient [52.2669490431145]
PropEn is inspired by'matching', which enables implicit guidance without training a discriminator.
We show that training with a matched dataset approximates the gradient of the property of interest while remaining within the data distribution.
arXiv Detail & Related papers (2024-05-28T11:30:19Z) - Diffusion Model for Data-Driven Black-Box Optimization [54.25693582870226]
We focus on diffusion models, a powerful generative AI technology, and investigate their potential for black-box optimization.
We study two practical types of labels: 1) noisy measurements of a real-valued reward function and 2) human preference based on pairwise comparisons.
Our proposed method reformulates the design optimization problem into a conditional sampling problem, which allows us to leverage the power of diffusion models.
arXiv Detail & Related papers (2024-03-20T00:41:12Z) - Functional Graphical Models: Structure Enables Offline Data-Driven Optimization [111.28605744661638]
We show how structure can enable sample-efficient data-driven optimization.
We also present a data-driven optimization algorithm that infers the FGM structure itself.
arXiv Detail & Related papers (2024-01-08T22:33:14Z) - Aligning Optimization Trajectories with Diffusion Models for Constrained
Design Generation [17.164961143132473]
We introduce a learning framework that demonstrates the efficacy of aligning the sampling trajectory of diffusion models with the optimization trajectory derived from traditional physics-based methods.
Our method allows for generating feasible and high-performance designs in as few as two steps without the need for expensive preprocessing, external surrogate models, or additional labeled data.
Our results demonstrate that TA outperforms state-of-the-art deep generative models on in-distribution configurations and halves the inference computational cost.
arXiv Detail & Related papers (2023-05-29T09:16:07Z) - Model-Based Deep Learning: On the Intersection of Deep Learning and
Optimization [101.32332941117271]
Decision making algorithms are used in a multitude of different applications.
Deep learning approaches that use highly parametric architectures tuned from data without relying on mathematical models are becoming increasingly popular.
Model-based optimization and data-centric deep learning are often considered to be distinct disciplines.
arXiv Detail & Related papers (2022-05-05T13:40:08Z) - Early-Phase Performance-Driven Design using Generative Models [0.0]
This research introduces a novel method for performance-driven geometry generation that can afford interaction directly in the 3d modeling environment.
The method uses Machine Learning techniques to train a generative model offline.
By navigating the generative model's latent space, geometries with the desired characteristics can be quickly generated.
arXiv Detail & Related papers (2021-07-19T01:25:11Z) - Good practices for Bayesian Optimization of high dimensional structured
spaces [15.488642552157131]
We study the effect of different search space design choices for performing Bayesian Optimization in high dimensional structured datasets.
We evaluate new methods to automatically define the optimization bounds in the latent space.
We provide recommendations for the practitioners.
arXiv Detail & Related papers (2020-12-31T07:00:39Z) - Optimization-Inspired Learning with Architecture Augmentations and
Control Mechanisms for Low-Level Vision [74.9260745577362]
This paper proposes a unified optimization-inspired learning framework to aggregate Generative, Discriminative, and Corrective (GDC) principles.
We construct three propagative modules to effectively solve the optimization models with flexible combinations.
Experiments across varied low-level vision tasks validate the efficacy and adaptability of GDC.
arXiv Detail & Related papers (2020-12-10T03:24:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.