Convergence analysis of unsupervised Legendre-Galerkin neural networks
for linear second-order elliptic PDEs
- URL: http://arxiv.org/abs/2211.08900v1
- Date: Wed, 16 Nov 2022 13:31:03 GMT
- Title: Convergence analysis of unsupervised Legendre-Galerkin neural networks
for linear second-order elliptic PDEs
- Authors: Seungchan Ko, Seok-Bae Yun and Youngjoon Hong
- Abstract summary: We perform the convergence analysis of unsupervised Legendre--Galerkin neural networks (ULGNet)
ULGNet is a deep-learning-based numerical method for solving partial differential equations (PDEs)
- Score: 0.8594140167290099
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this paper, we perform the convergence analysis of unsupervised
Legendre--Galerkin neural networks (ULGNet), a deep-learning-based numerical
method for solving partial differential equations (PDEs). Unlike existing deep
learning-based numerical methods for PDEs, the ULGNet expresses the solution as
a spectral expansion with respect to the Legendre basis and predicts the
coefficients with deep neural networks by solving a variational residual
minimization problem. Since the corresponding loss function is equivalent to
the residual induced by the linear algebraic system depending on the choice of
basis functions, we prove that the minimizer of the discrete loss function
converges to the weak solution of the PDEs. Numerical evidence will also be
provided to support the theoretical result. Key technical tools include the
variant of the universal approximation theorem for bounded neural networks, the
analysis of the stiffness and mass matrices, and the uniform law of large
numbers in terms of the Rademacher complexity.
Related papers
- First Order System Least Squares Neural Networks [0.0]
We numerically solve PDEs on bounded, polytopal domains in euclidean spaces by deep neural networks.
An adaptive neural network growth strategy is proposed which, assuming exact numerical minimization of the LSQ loss functional, yields sequences of neural networks with realizations that converge rate-optimally to the exact solution of the first order system LSQ formulation.
arXiv Detail & Related papers (2024-09-30T13:04:35Z) - Neural networks for bifurcation and linear stability analysis of steady states in partial differential equations [0.0]
A neural network is proposed to construct bifurcation diagrams from parameterized nonlinear PDEs.
A neural network approach is also presented for solving eigenvalue problems to analyze solution linear stability.
arXiv Detail & Related papers (2024-07-29T05:05:13Z) - Lie Point Symmetry and Physics Informed Networks [59.56218517113066]
We propose a loss function that informs the network about Lie point symmetries in the same way that PINN models try to enforce the underlying PDE through a loss function.
Our symmetry loss ensures that the infinitesimal generators of the Lie group conserve the PDE solutions.
Empirical evaluations indicate that the inductive bias introduced by the Lie point symmetries of the PDEs greatly boosts the sample efficiency of PINNs.
arXiv Detail & Related papers (2023-11-07T19:07:16Z) - Learning Discretized Neural Networks under Ricci Flow [51.36292559262042]
We study Discretized Neural Networks (DNNs) composed of low-precision weights and activations.
DNNs suffer from either infinite or zero gradients due to the non-differentiable discrete function during training.
arXiv Detail & Related papers (2023-02-07T10:51:53Z) - Tunable Complexity Benchmarks for Evaluating Physics-Informed Neural
Networks on Coupled Ordinary Differential Equations [64.78260098263489]
In this work, we assess the ability of physics-informed neural networks (PINNs) to solve increasingly-complex coupled ordinary differential equations (ODEs)
We show that PINNs eventually fail to produce correct solutions to these benchmarks as their complexity increases.
We identify several reasons why this may be the case, including insufficient network capacity, poor conditioning of the ODEs, and high local curvature, as measured by the Laplacian of the PINN loss.
arXiv Detail & Related papers (2022-10-14T15:01:32Z) - Mean-field Analysis of Piecewise Linear Solutions for Wide ReLU Networks [83.58049517083138]
We consider a two-layer ReLU network trained via gradient descent.
We show that SGD is biased towards a simple solution.
We also provide empirical evidence that knots at locations distinct from the data points might occur.
arXiv Detail & Related papers (2021-11-03T15:14:20Z) - DiffNet: Neural Field Solutions of Parametric Partial Differential
Equations [30.80582606420882]
We consider a mesh-based approach for training a neural network to produce field predictions of solutions to PDEs.
We use a weighted Galerkin loss function based on the Finite Element Method (FEM) on a parametric elliptic PDE.
We prove theoretically, and illustrate with experiments, convergence results analogous to mesh convergence analysis deployed in finite element solutions to PDEs.
arXiv Detail & Related papers (2021-10-04T17:59:18Z) - Solving PDEs on Unknown Manifolds with Machine Learning [8.220217498103315]
This paper presents a mesh-free computational framework and machine learning theory for solving elliptic PDEs on unknown manifold.
We show that the proposed NN solver can robustly generalize the PDE on new data points with errors that are almost identical to generalizations on new data points.
arXiv Detail & Related papers (2021-06-12T03:55:15Z) - Parametric Complexity Bounds for Approximating PDEs with Neural Networks [41.46028070204925]
We prove that when a PDE's coefficients are representable by small neural networks, the parameters required to approximate its solution scalely with the input $d$ are proportional to the parameter counts of the neural networks.
Our proof is based on constructing a neural network which simulates gradient descent in an appropriate space which converges to the solution of the PDE.
arXiv Detail & Related papers (2021-03-03T02:42:57Z) - Provably Efficient Neural Estimation of Structural Equation Model: An
Adversarial Approach [144.21892195917758]
We study estimation in a class of generalized Structural equation models (SEMs)
We formulate the linear operator equation as a min-max game, where both players are parameterized by neural networks (NNs), and learn the parameters of these neural networks using a gradient descent.
For the first time we provide a tractable estimation procedure for SEMs based on NNs with provable convergence and without the need for sample splitting.
arXiv Detail & Related papers (2020-07-02T17:55:47Z) - Multipole Graph Neural Operator for Parametric Partial Differential
Equations [57.90284928158383]
One of the main challenges in using deep learning-based methods for simulating physical systems is formulating physics-based data.
We propose a novel multi-level graph neural network framework that captures interaction at all ranges with only linear complexity.
Experiments confirm our multi-graph network learns discretization-invariant solution operators to PDEs and can be evaluated in linear time.
arXiv Detail & Related papers (2020-06-16T21:56:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.