Deep neural network surrogates for non-smooth quantities of interest in
shape uncertainty quantification
- URL: http://arxiv.org/abs/2101.07023v1
- Date: Mon, 18 Jan 2021 12:02:57 GMT
- Title: Deep neural network surrogates for non-smooth quantities of interest in
shape uncertainty quantification
- Authors: Laura Scarabosio
- Abstract summary: We focus on an elliptic interface problem and a Helmholtz transmission problem.
Point values of the solution in the physical domain depend in general non-smoothly on the high-dimensional parameter.
We build surrogates for point evaluation using deep neural networks.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We consider the point evaluation of the solution to interface problems with
geometric uncertainties, where the uncertainty in the obstacle is described by
a high-dimensional parameter $\boldsymbol{y}\in[-1,1]^d$, $d\in\mathbb{N}$. We
focus in particular on an elliptic interface problem and a Helmholtz
transmission problem. Point values of the solution in the physical domain
depend in general non-smoothly on the high-dimensional parameter, posing a
challenge when one is interested in building surrogates. Indeed, high-order
methods show poor convergence rates, while methods which are able to track
discontinuities usually suffer from the so-called curse of dimensionality. For
this reason, in this work we propose to build surrogates for point evaluation
using deep neural networks. We provide a theoretical justification for why we
expect neural networks to provide good surrogates. Furthermore, we present
extensive numerical experiments showing their good performance in practice. We
observe in particular that neural networks do not suffer from the curse of
dimensionality, and we study the dependence of the error on the number of point
evaluations (that is, the number of discontinuities in the parameter space), as
well as on several modeling parameters, such as the contrast between the two
materials and, for the Helmholtz transmission problem, the wavenumber.
Related papers
- Neural network analysis of neutron and X-ray reflectivity data:
Incorporating prior knowledge for tackling the phase problem [141.5628276096321]
We present an approach that utilizes prior knowledge to regularize the training process over larger parameter spaces.
We demonstrate the effectiveness of our method in various scenarios, including multilayer structures with box model parameterization.
In contrast to previous methods, our approach scales favorably when increasing the complexity of the inverse problem.
arXiv Detail & Related papers (2023-06-28T11:15:53Z) - On the ISS Property of the Gradient Flow for Single Hidden-Layer Neural
Networks with Linear Activations [0.0]
We investigate the effects of overfitting on the robustness of gradient-descent training when subject to uncertainty on the gradient estimation.
We show that the general overparametrized formulation introduces a set of spurious equilibria which lay outside the set where the loss function is minimized.
arXiv Detail & Related papers (2023-05-17T02:26:34Z) - Conductivity Imaging from Internal Measurements with Mixed Least-Squares
Deep Neural Networks [4.228167013618626]
We develop a novel approach using deep neural networks to reconstruct the conductivity distribution in elliptic problems.
We provide a thorough analysis of the deep neural network approximations of the conductivity for both continuous and empirical losses.
arXiv Detail & Related papers (2023-03-29T04:43:03Z) - Efficient Bayesian Physics Informed Neural Networks for Inverse Problems
via Ensemble Kalman Inversion [0.0]
We present a new efficient inference algorithm for B-PINNs that uses Ensemble Kalman Inversion (EKI) for high-dimensional inference tasks.
We find that our proposed method can achieve inference results with informative uncertainty estimates comparable to Hamiltonian Monte Carlo (HMC)-based B-PINNs with a much reduced computational cost.
arXiv Detail & Related papers (2023-03-13T18:15:26Z) - Sample Complexity of Nonparametric Off-Policy Evaluation on
Low-Dimensional Manifolds using Deep Networks [71.95722100511627]
We consider the off-policy evaluation problem of reinforcement learning using deep neural networks.
We show that, by choosing network size appropriately, one can leverage the low-dimensional manifold structure in the Markov decision process.
arXiv Detail & Related papers (2022-06-06T20:25:20Z) - Mean-field Analysis of Piecewise Linear Solutions for Wide ReLU Networks [83.58049517083138]
We consider a two-layer ReLU network trained via gradient descent.
We show that SGD is biased towards a simple solution.
We also provide empirical evidence that knots at locations distinct from the data points might occur.
arXiv Detail & Related papers (2021-11-03T15:14:20Z) - And/or trade-off in artificial neurons: impact on adversarial robustness [91.3755431537592]
Presence of sufficient number of OR-like neurons in a network can lead to classification brittleness and increased vulnerability to adversarial attacks.
We define AND-like neurons and propose measures to increase their proportion in the network.
Experimental results on the MNIST dataset suggest that our approach holds promise as a direction for further exploration.
arXiv Detail & Related papers (2021-02-15T08:19:05Z) - Attribute-Guided Adversarial Training for Robustness to Natural
Perturbations [64.35805267250682]
We propose an adversarial training approach which learns to generate new samples so as to maximize exposure of the classifier to the attributes-space.
Our approach enables deep neural networks to be robust against a wide range of naturally occurring perturbations.
arXiv Detail & Related papers (2020-12-03T10:17:30Z) - Numerical Solution of Inverse Problems by Weak Adversarial Networks [6.571303769953873]
We leverage the weak formulation of PDE in the given inverse problem, and parameterize the solution and the test function as deep neural networks.
As the parameters are updated, the network gradually approximates the solution of the inverse problem.
arXiv Detail & Related papers (2020-02-26T07:58:37Z) - Beyond Dropout: Feature Map Distortion to Regularize Deep Neural
Networks [107.77595511218429]
In this paper, we investigate the empirical Rademacher complexity related to intermediate layers of deep neural networks.
We propose a feature distortion method (Disout) for addressing the aforementioned problem.
The superiority of the proposed feature map distortion for producing deep neural network with higher testing performance is analyzed and demonstrated.
arXiv Detail & Related papers (2020-02-23T13:59:13Z) - Properties of the geometry of solutions and capacity of multi-layer neural networks with Rectified Linear Units activations [2.3018169548556977]
We study the effects of Rectified Linear Units on the capacity and on the geometrical landscape of the solution space in two-layer neural networks.
We find that, quite unexpectedly, the capacity of the network remains finite as the number of neurons in the hidden layer increases.
Possibly more important, a large deviation approach allows us to find that the geometrical landscape of the solution space has a peculiar structure.
arXiv Detail & Related papers (2019-07-17T15:23:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.