NITO: Neural Implicit Fields for Resolution-free Topology Optimization
- URL: http://arxiv.org/abs/2402.05073v1
- Date: Wed, 7 Feb 2024 18:27:29 GMT
- Title: NITO: Neural Implicit Fields for Resolution-free Topology Optimization
- Authors: Amin Heyrani Nobari, Giorgio Giannone, Lyle Regenwetter, Faez Ahmed
- Abstract summary: Topology optimization is a critical task in engineering design, where the goal is to optimally distribute material in a given space.
We introduce Neural Implicit Topology Optimization (NITO), a novel approach to accelerate topology optimization problems using deep learning.
NITO synthesizes structures with up to seven times better structural efficiency compared to SOTA diffusion models.
- Score: 7.338114424386579
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Topology optimization is a critical task in engineering design, where the
goal is to optimally distribute material in a given space for maximum
performance. We introduce Neural Implicit Topology Optimization (NITO), a novel
approach to accelerate topology optimization problems using deep learning. NITO
stands out as one of the first frameworks to offer a resolution-free and
domain-agnostic solution in deep learning-based topology optimization. NITO
synthesizes structures with up to seven times better structural efficiency
compared to SOTA diffusion models and does so in a tenth of the time. In the
NITO framework, we introduce a novel method, the Boundary Point Order-Invariant
MLP (BPOM), to represent boundary conditions in a sparse and domain-agnostic
manner, moving away from expensive simulation-based approaches. Crucially, NITO
circumvents the domain and resolution limitations that restrict Convolutional
Neural Network (CNN) models to a structured domain of fixed size -- limitations
that hinder the widespread adoption of CNNs in engineering applications. This
generalizability allows a single NITO model to train and generate solutions in
countless domains, eliminating the need for numerous domain-specific CNNs and
their extensive datasets. Despite its generalizability, NITO outperforms SOTA
models even in specialized tasks, is an order of magnitude smaller, and is
practically trainable at high resolutions that would be restrictive for CNNs.
This combination of versatility, efficiency, and performance underlines NITO's
potential to transform the landscape of engineering design optimization
problems through implicit fields.
Related papers
- N-Adaptive Ritz Method: A Neural Network Enriched Partition of Unity for
Boundary Value Problems [1.2200609701777907]
This work introduces a novel neural network-enriched Partition of Unity (NN-PU) approach for solving boundary value problems via artificial neural networks.
The NN enrichment is constructed by combining pre-trained feature-encoded NN blocks with an untrained NN block.
The proposed method offers accurate solutions while notably reducing the computational cost compared to the conventional adaptive refinement in the mesh-based methods.
arXiv Detail & Related papers (2024-01-16T18:11:14Z) - Achieving Constraints in Neural Networks: A Stochastic Augmented
Lagrangian Approach [49.1574468325115]
Regularizing Deep Neural Networks (DNNs) is essential for improving generalizability and preventing overfitting.
We propose a novel approach to DNN regularization by framing the training process as a constrained optimization problem.
We employ the Augmented Lagrangian (SAL) method to achieve a more flexible and efficient regularization mechanism.
arXiv Detail & Related papers (2023-10-25T13:55:35Z) - Neural Fields with Hard Constraints of Arbitrary Differential Order [61.49418682745144]
We develop a series of approaches for enforcing hard constraints on neural fields.
The constraints can be specified as a linear operator applied to the neural field and its derivatives.
Our approaches are demonstrated in a wide range of real-world applications.
arXiv Detail & Related papers (2023-06-15T08:33:52Z) - DMF-TONN: Direct Mesh-free Topology Optimization using Neural Networks [4.663709549795511]
We propose a direct mesh-free method for performing topology optimization by integrating a density field approximation neural network with a displacement field approximation neural network.
We show that this direct integration approach can give comparable results to conventional topology optimization techniques.
arXiv Detail & Related papers (2023-05-06T18:04:51Z) - A mechanistic-based data-driven approach to accelerate structural
topology optimization through finite element convolutional neural network
(FE-CNN) [5.469226380238751]
A mechanistic data-driven approach is proposed to accelerate structural topology optimization.
Our approach can be divided into two stages: offline training, and online optimization.
Numerical examples demonstrate that this approach can accelerate optimization by up to an order of magnitude in computational time.
arXiv Detail & Related papers (2021-06-25T14:11:45Z) - NTopo: Mesh-free Topology Optimization using Implicit Neural
Representations [35.07884509198916]
We present a novel machine learning approach to tackle topology optimization problems.
We use multilayer perceptrons (MLPs) to parameterize both density and displacement fields.
As we show through our experiments, a major benefit of our approach is that it enables self-supervised learning of continuous solution spaces.
arXiv Detail & Related papers (2021-02-22T05:25:22Z) - Offline Model-Based Optimization via Normalized Maximum Likelihood
Estimation [101.22379613810881]
We consider data-driven optimization problems where one must maximize a function given only queries at a fixed set of points.
This problem setting emerges in many domains where function evaluation is a complex and expensive process.
We propose a tractable approximation that allows us to scale our method to high-capacity neural network models.
arXiv Detail & Related papers (2021-02-16T06:04:27Z) - Modeling from Features: a Mean-field Framework for Over-parameterized
Deep Neural Networks [54.27962244835622]
This paper proposes a new mean-field framework for over- parameterized deep neural networks (DNNs)
In this framework, a DNN is represented by probability measures and functions over its features in the continuous limit.
We illustrate the framework via the standard DNN and the Residual Network (Res-Net) architectures.
arXiv Detail & Related papers (2020-07-03T01:37:16Z) - Self-Directed Online Machine Learning for Topology Optimization [58.920693413667216]
Self-directed Online Learning Optimization integrates Deep Neural Network (DNN) with Finite Element Method (FEM) calculations.
Our algorithm was tested by four types of problems including compliance minimization, fluid-structure optimization, heat transfer enhancement and truss optimization.
It reduced the computational time by 2 5 orders of magnitude compared with directly using methods, and outperformed all state-of-the-art algorithms tested in our experiments.
arXiv Detail & Related papers (2020-02-04T20:00:28Z) - Optimizing Wireless Systems Using Unsupervised and
Reinforced-Unsupervised Deep Learning [96.01176486957226]
Resource allocation and transceivers in wireless networks are usually designed by solving optimization problems.
In this article, we introduce unsupervised and reinforced-unsupervised learning frameworks for solving both variable and functional optimization problems.
arXiv Detail & Related papers (2020-01-03T11:01:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.