On the convergence of physics informed neural networks for linear
second-order elliptic and parabolic type PDEs
- URL: http://arxiv.org/abs/2004.01806v2
- Date: Wed, 21 Oct 2020 19:16:36 GMT
- Title: On the convergence of physics informed neural networks for linear
second-order elliptic and parabolic type PDEs
- Authors: Yeonjong Shin, Jerome Darbon, George Em Karniadakis
- Abstract summary: Physics informed neural networks (PINNs) are deep learning based techniques for solving partial differential equations (PDEs)
We show that the sequence of minimizers strongly converges to the PDE solution in $C0$.
To the best of our knowledge, this is the first theoretical work that shows the consistency of PINNs.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Physics informed neural networks (PINNs) are deep learning based techniques
for solving partial differential equations (PDEs) encounted in computational
science and engineering. Guided by data and physical laws, PINNs find a neural
network that approximates the solution to a system of PDEs. Such a neural
network is obtained by minimizing a loss function in which any prior knowledge
of PDEs and data are encoded. Despite its remarkable empirical success in one,
two or three dimensional problems, there is little theoretical justification
for PINNs.
As the number of data grows, PINNs generate a sequence of minimizers which
correspond to a sequence of neural networks. We want to answer the question:
Does the sequence of minimizers converge to the solution to the PDE? We
consider two classes of PDEs: linear second-order elliptic and parabolic. By
adapting the Schauder approach and the maximum principle, we show that the
sequence of minimizers strongly converges to the PDE solution in $C^0$.
Furthermore, we show that if each minimizer satisfies the initial/boundary
conditions, the convergence mode becomes $H^1$. Computational examples are
provided to illustrate our theoretical findings. To the best of our knowledge,
this is the first theoretical work that shows the consistency of PINNs.
Related papers
- Improving PINNs By Algebraic Inclusion of Boundary and Initial Conditions [0.1874930567916036]
"AI for Science" aims to solve fundamental scientific problems using AI techniques.
In this work we explore the possibility of changing the model being trained from being just a neural network to being a non-linear transformation of it.
This reduces the number of terms in the loss function than the standard PINN losses.
arXiv Detail & Related papers (2024-07-30T11:19:48Z) - Global Convergence of Deep Galerkin and PINNs Methods for Solving
Partial Differential Equations [0.0]
A variety of deep learning methods have been developed to try and solve high-dimensional PDEs by approximating the solution using a neural network.
We prove global convergence for one of the commonly-used deep learning algorithms for solving PDEs, the Deep Galerkin MethodDGM.
arXiv Detail & Related papers (2023-05-10T09:20:11Z) - iPINNs: Incremental learning for Physics-informed neural networks [66.4795381419701]
Physics-informed neural networks (PINNs) have recently become a powerful tool for solving partial differential equations (PDEs)
We propose incremental PINNs that can learn multiple tasks sequentially without additional parameters for new tasks and improve performance for every equation in the sequence.
Our approach learns multiple PDEs starting from the simplest one by creating its own subnetwork for each PDE and allowing each subnetwork to overlap with previously learnedworks.
arXiv Detail & Related papers (2023-04-10T20:19:20Z) - Learning to Solve PDE-constrained Inverse Problems with Graph Networks [51.89325993156204]
In many application domains across science and engineering, we are interested in solving inverse problems with constraints defined by a partial differential equation (PDE)
Here we explore GNNs to solve such PDE-constrained inverse problems.
We demonstrate computational speedups of up to 90x using GNNs compared to principled solvers.
arXiv Detail & Related papers (2022-06-01T18:48:01Z) - Improved Training of Physics-Informed Neural Networks with Model
Ensembles [81.38804205212425]
We propose to expand the solution interval gradually to make the PINN converge to the correct solution.
All ensemble members converge to the same solution in the vicinity of observed data.
We show experimentally that the proposed method can improve the accuracy of the found solution.
arXiv Detail & Related papers (2022-04-11T14:05:34Z) - Learning Physics-Informed Neural Networks without Stacked
Back-propagation [82.26566759276105]
We develop a novel approach that can significantly accelerate the training of Physics-Informed Neural Networks.
In particular, we parameterize the PDE solution by the Gaussian smoothed model and show that, derived from Stein's Identity, the second-order derivatives can be efficiently calculated without back-propagation.
Experimental results show that our proposed method can achieve competitive error compared to standard PINN training but is two orders of magnitude faster.
arXiv Detail & Related papers (2022-02-18T18:07:54Z) - Characterizing possible failure modes in physics-informed neural
networks [55.83255669840384]
Recent work in scientific machine learning has developed so-called physics-informed neural network (PINN) models.
We demonstrate that, while existing PINN methodologies can learn good models for relatively trivial problems, they can easily fail to learn relevant physical phenomena even for simple PDEs.
We show that these possible failure modes are not due to the lack of expressivity in the NN architecture, but that the PINN's setup makes the loss landscape very hard to optimize.
arXiv Detail & Related papers (2021-09-02T16:06:45Z) - Solving PDEs on Unknown Manifolds with Machine Learning [8.220217498103315]
This paper presents a mesh-free computational framework and machine learning theory for solving elliptic PDEs on unknown manifold.
We show that the proposed NN solver can robustly generalize the PDE on new data points with errors that are almost identical to generalizations on new data points.
arXiv Detail & Related papers (2021-06-12T03:55:15Z) - dNNsolve: an efficient NN-based PDE solver [62.997667081978825]
We introduce dNNsolve, that makes use of dual Neural Networks to solve ODEs/PDEs.
We show that dNNsolve is capable of solving a broad range of ODEs/PDEs in 1, 2 and 3 spacetime dimensions.
arXiv Detail & Related papers (2021-03-15T19:14:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.