Concurrent build direction, part segmentation, and topology optimization
for additive manufacturing using neural networks
- URL: http://arxiv.org/abs/2210.01315v1
- Date: Tue, 4 Oct 2022 02:17:54 GMT
- Title: Concurrent build direction, part segmentation, and topology optimization
for additive manufacturing using neural networks
- Authors: Hongrui Chen, Aditya Joglekar, Kate S. Whitefoot, Levent Burak Kara
- Abstract summary: We propose a neural network-based approach to topology optimization that aims to reduce the use of support structures in additive manufacturing.
Our approach uses a network architecture that allows the simultaneous determination of an optimized: (1) part segmentation, (2) the topology of each part, and (3) the build direction of each part.
- Score: 2.2911466677853065
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: We propose a neural network-based approach to topology optimization that aims
to reduce the use of support structures in additive manufacturing. Our approach
uses a network architecture that allows the simultaneous determination of an
optimized: (1) part segmentation, (2) the topology of each part, and (3) the
build direction of each part that collectively minimize the amount of support
structure. Through training, the network learns a material density and segment
classification in the continuous 3D space. Given a problem domain with
prescribed load and displacement boundary conditions, the neural network takes
as input 3D coordinates of the voxelized domain as training samples and outputs
a continuous density field. Since the neural network for topology optimization
learns the density distribution field, analytical solutions to the density
gradient can be obtained from the input-output relationship of the neural
network. We demonstrate our approach on several compliance minimization
problems with volume fraction constraints, where support volume minimization is
added as an additional criterion to the objective function. We show that
simultaneous optimization of part segmentation along with the topology and
print angle optimization further reduces the support structure, compared to a
combined print angle and topology optimization without segmentation.
Related papers
- From Complexity to Clarity: Analytical Expressions of Deep Neural Network Weights via Clifford's Geometric Algebra and Convexity [54.01594785269913]
We show that optimal weights of deep ReLU neural networks are given by the wedge product of training samples when trained with standard regularized loss.
The training problem reduces to convex optimization over wedge product features, which encode the geometric structure of the training dataset.
arXiv Detail & Related papers (2023-09-28T15:19:30Z) - Large-scale global optimization of ultra-high dimensional non-convex
landscapes based on generative neural networks [0.0]
We present an algorithm manage ultra-high dimensional optimization.
based on a deep generative network.
We show that our method performs better with fewer function evaluations compared to state-of-the-art algorithm.
arXiv Detail & Related papers (2023-07-09T00:05:59Z) - Topology Optimization using Neural Networks with Conditioning Field
Initialization for Improved Efficiency [2.575019247624295]
We show that by using a prior initial field on the unoptimized domain, the efficiency of neural network based topology optimization can be further improved.
We employ the strain energy field calculated on the initial design domain as an additional conditioning field input to the neural network throughout the optimization.
arXiv Detail & Related papers (2023-05-17T07:42:24Z) - DMF-TONN: Direct Mesh-free Topology Optimization using Neural Networks [4.663709549795511]
We propose a direct mesh-free method for performing topology optimization by integrating a density field approximation neural network with a displacement field approximation neural network.
We show that this direct integration approach can give comparable results to conventional topology optimization techniques.
arXiv Detail & Related papers (2023-05-06T18:04:51Z) - Mean-field Analysis of Piecewise Linear Solutions for Wide ReLU Networks [83.58049517083138]
We consider a two-layer ReLU network trained via gradient descent.
We show that SGD is biased towards a simple solution.
We also provide empirical evidence that knots at locations distinct from the data points might occur.
arXiv Detail & Related papers (2021-11-03T15:14:20Z) - De-homogenization using Convolutional Neural Networks [1.0323063834827415]
This paper presents a deep learning-based de-homogenization method for structural compliance minimization.
For an appropriate choice of parameters, the de-homogenized designs perform within $7-25%$ of the homogenization-based solution.
arXiv Detail & Related papers (2021-05-10T09:50:06Z) - InverseForm: A Loss Function for Structured Boundary-Aware Segmentation [80.39674800972182]
We present a novel boundary-aware loss term for semantic segmentation using an inverse-transformation network.
This plug-in loss term complements the cross-entropy loss in capturing boundary transformations.
We analyze the quantitative and qualitative effects of our loss function on three indoor and outdoor segmentation benchmarks.
arXiv Detail & Related papers (2021-04-06T18:52:45Z) - Modeling from Features: a Mean-field Framework for Over-parameterized
Deep Neural Networks [54.27962244835622]
This paper proposes a new mean-field framework for over- parameterized deep neural networks (DNNs)
In this framework, a DNN is represented by probability measures and functions over its features in the continuous limit.
We illustrate the framework via the standard DNN and the Residual Network (Res-Net) architectures.
arXiv Detail & Related papers (2020-07-03T01:37:16Z) - The Hidden Convex Optimization Landscape of Two-Layer ReLU Neural
Networks: an Exact Characterization of the Optimal Solutions [51.60996023961886]
We prove that finding all globally optimal two-layer ReLU neural networks can be performed by solving a convex optimization program with cone constraints.
Our analysis is novel, characterizes all optimal solutions, and does not leverage duality-based analysis which was recently used to lift neural network training into convex spaces.
arXiv Detail & Related papers (2020-06-10T15:38:30Z) - Neural Subdivision [58.97214948753937]
This paper introduces Neural Subdivision, a novel framework for data-driven coarseto-fine geometry modeling.
We optimize for the same set of network weights across all local mesh patches, thus providing an architecture that is not constrained to a specific input mesh, fixed genus, or category.
We demonstrate that even when trained on a single high-resolution mesh our method generates reasonable subdivisions for novel shapes.
arXiv Detail & Related papers (2020-05-04T20:03:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.