Critical Investigation of Failure Modes in Physics-informed Neural
Networks
- URL: http://arxiv.org/abs/2206.09961v1
- Date: Mon, 20 Jun 2022 18:43:35 GMT
- Title: Critical Investigation of Failure Modes in Physics-informed Neural
Networks
- Authors: Shamsulhaq Basir, Inanc Senocak
- Abstract summary: We show that a physics-informed neural network with a composite formulation produces highly non- learned loss surfaces that are difficult to optimize.
We also assess the training both approaches on two elliptic problems with increasingly complex target solutions.
- Score: 0.9137554315375919
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Several recent works in scientific machine learning have revived interest in
the application of neural networks to partial differential equations (PDEs). A
popular approach is to aggregate the residual form of the governing PDE and its
boundary conditions as soft penalties into a composite objective/loss function
for training neural networks, which is commonly referred to as physics-informed
neural networks (PINNs). In the present study, we visualize the loss landscapes
and distributions of learned parameters and explain the ways this particular
formulation of the objective function may hinder or even prevent convergence
when dealing with challenging target solutions. We construct a purely
data-driven loss function composed of both the boundary loss and the domain
loss. Using this data-driven loss function and, separately, a physics-informed
loss function, we then train two neural network models with the same
architecture. We show that incomparable scales between boundary and domain loss
terms are the culprit behind the poor performance. Additionally, we assess the
performance of both approaches on two elliptic problems with increasingly
complex target solutions. Based on our analysis of their loss landscapes and
learned parameter distributions, we observe that a physics-informed neural
network with a composite objective function formulation produces highly
non-convex loss surfaces that are difficult to optimize and are more prone to
the problem of vanishing gradients.
Related papers
- Unraveling the Hessian: A Key to Smooth Convergence in Loss Function Landscapes [0.0]
We theoretically analyze the convergence of the loss landscape in a fully connected neural network and derive upper bounds for the difference in loss function values when adding a new object to the sample.
Our empirical study confirms these results on various datasets, demonstrating the convergence of the loss function surface for image classification tasks.
arXiv Detail & Related papers (2024-09-18T14:04:15Z) - Data-Guided Physics-Informed Neural Networks for Solving Inverse Problems in Partial Differential Equations [3.231578125191615]
Physics-informed neural networks (PINNs) represent a significant advancement in machine learning.
PINNs have been successfully applied to solve various forward and inverse problems in partial differential equations (PDEs)
This study proposes a novel framework termed data-guided physics-informed neural networks (DG-PINNs)
arXiv Detail & Related papers (2024-07-15T15:47:24Z) - Loss Jump During Loss Switch in Solving PDEs with Neural Networks [11.123662745891677]
Using neural networks to solve partial differential equations (PDEs) is gaining popularity as an alternative approach in the scientific computing community.
This work focuses on investigating how different loss functions impact the training of neural networks for solving PDEs.
arXiv Detail & Related papers (2024-05-06T01:18:36Z) - A topological description of loss surfaces based on Betti Numbers [8.539445673580252]
We provide a topological measure to evaluate loss complexity in the case of multilayer neural networks.
We find that certain variations in the loss function or model architecture, such as adding an $ell$ regularization term or skip connections in a feedforward network, do not affect loss in specific cases.
arXiv Detail & Related papers (2024-01-08T11:20:04Z) - Implicit Stochastic Gradient Descent for Training Physics-informed
Neural Networks [51.92362217307946]
Physics-informed neural networks (PINNs) have effectively been demonstrated in solving forward and inverse differential equation problems.
PINNs are trapped in training failures when the target functions to be approximated exhibit high-frequency or multi-scale features.
In this paper, we propose to employ implicit gradient descent (ISGD) method to train PINNs for improving the stability of training process.
arXiv Detail & Related papers (2023-03-03T08:17:47Z) - NeuralStagger: Accelerating Physics-constrained Neural PDE Solver with
Spatial-temporal Decomposition [67.46012350241969]
This paper proposes a general acceleration methodology called NeuralStagger.
It decomposing the original learning tasks into several coarser-resolution subtasks.
We demonstrate the successful application of NeuralStagger on 2D and 3D fluid dynamics simulations.
arXiv Detail & Related papers (2023-02-20T19:36:52Z) - Physics informed neural networks for continuum micromechanics [68.8204255655161]
Recently, physics informed neural networks have successfully been applied to a broad variety of problems in applied mathematics and engineering.
Due to the global approximation, physics informed neural networks have difficulties in displaying localized effects and strong non-linear solutions by optimization.
It is shown, that the domain decomposition approach is able to accurately resolve nonlinear stress, displacement and energy fields in heterogeneous microstructures obtained from real-world $mu$CT-scans.
arXiv Detail & Related papers (2021-10-14T14:05:19Z) - Characterizing possible failure modes in physics-informed neural
networks [55.83255669840384]
Recent work in scientific machine learning has developed so-called physics-informed neural network (PINN) models.
We demonstrate that, while existing PINN methodologies can learn good models for relatively trivial problems, they can easily fail to learn relevant physical phenomena even for simple PDEs.
We show that these possible failure modes are not due to the lack of expressivity in the NN architecture, but that the PINN's setup makes the loss landscape very hard to optimize.
arXiv Detail & Related papers (2021-09-02T16:06:45Z) - Topological obstructions in neural networks learning [67.8848058842671]
We study global properties of the loss gradient function flow.
We use topological data analysis of the loss function and its Morse complex to relate local behavior along gradient trajectories with global properties of the loss surface.
arXiv Detail & Related papers (2020-12-31T18:53:25Z) - Binary Neural Networks: A Survey [126.67799882857656]
The binary neural network serves as a promising technique for deploying deep models on resource-limited devices.
The binarization inevitably causes severe information loss, and even worse, its discontinuity brings difficulty to the optimization of the deep network.
We present a survey of these algorithms, mainly categorized into the native solutions directly conducting binarization, and the optimized ones using techniques like minimizing the quantization error, improving the network loss function, and reducing the gradient error.
arXiv Detail & Related papers (2020-03-31T16:47:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.