Local Extreme Learning Machines and Domain Decomposition for Solving
Linear and Nonlinear Partial Differential Equations
- URL: http://arxiv.org/abs/2012.02895v1
- Date: Fri, 4 Dec 2020 23:19:39 GMT
- Title: Local Extreme Learning Machines and Domain Decomposition for Solving
Linear and Nonlinear Partial Differential Equations
- Authors: Suchuan Dong, Zongwei Li
- Abstract summary: We present a neural network-based method for solving linear and nonlinear partial differential equations.
The method combines the ideas of extreme learning machines (ELM), domain decomposition and local neural networks.
We compare the current method with the deep Galerkin method (DGM) and the physics-informed neural network (PINN) in terms of the accuracy and computational cost.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present a neural network-based method for solving linear and nonlinear
partial differential equations, by combining the ideas of extreme learning
machines (ELM), domain decomposition and local neural networks. The field
solution on each sub-domain is represented by a local feed-forward neural
network, and $C^k$ continuity is imposed on the sub-domain boundaries. Each
local neural network consists of a small number of hidden layers, while its
last hidden layer can be wide. The weight/bias coefficients in all hidden
layers of the local neural networks are pre-set to random values and are fixed,
and only the weight coefficients in the output layers are training parameters.
The overall neural network is trained by a linear or nonlinear least squares
computation, not by the back-propagation type algorithms. We introduce a block
time-marching scheme together with the presented method for long-time dynamic
simulations. The current method exhibits a clear sense of convergence with
respect to the degrees of freedom in the neural network. Its numerical errors
typically decrease exponentially or nearly exponentially as the number of
degrees of freedom increases. Extensive numerical experiments have been
performed to demonstrate the computational performance of the presented method.
We compare the current method with the deep Galerkin method (DGM) and the
physics-informed neural network (PINN) in terms of the accuracy and
computational cost. The current method exhibits a clear superiority, with its
numerical errors and network training time considerably smaller (typically by
orders of magnitude) than those of DGM and PINN. We also compare the current
method with the classical finite element method (FEM). The computational
performance of the current method is on par with, and oftentimes exceeds, the
FEM performance.
Related papers
- A Nonoverlapping Domain Decomposition Method for Extreme Learning Machines: Elliptic Problems [0.0]
Extreme learning machine (ELM) is a methodology for solving partial differential equations (PDEs) using a single hidden layer feed-forward neural network.
In this paper, we propose a nonoverlapping domain decomposition method (DDM) for ELMs that not only reduces the training time of ELMs, but is also suitable for parallel computation.
arXiv Detail & Related papers (2024-06-22T23:25:54Z) - Implicit regularization of deep residual networks towards neural ODEs [8.075122862553359]
We establish an implicit regularization of deep residual networks towards neural ODEs.
We prove that if the network is as a discretization of a neural ODE, then such a discretization holds throughout training.
arXiv Detail & Related papers (2023-09-03T16:35:59Z) - Speed Limits for Deep Learning [67.69149326107103]
Recent advancement in thermodynamics allows bounding the speed at which one can go from the initial weight distribution to the final distribution of the fully trained network.
We provide analytical expressions for these speed limits for linear and linearizable neural networks.
Remarkably, given some plausible scaling assumptions on the NTK spectra and spectral decomposition of the labels -- learning is optimal in a scaling sense.
arXiv Detail & Related papers (2023-07-27T06:59:46Z) - Globally Optimal Training of Neural Networks with Threshold Activation
Functions [63.03759813952481]
We study weight decay regularized training problems of deep neural networks with threshold activations.
We derive a simplified convex optimization formulation when the dataset can be shattered at a certain layer of the network.
arXiv Detail & Related papers (2023-03-06T18:59:13Z) - Local neural operator for solving transient partial differential
equations on varied domains [8.905324065830861]
We propose local neural operator (LNO) for solving transient partial differential equations (PDEs) on varied domains.
It comes together with a handy strategy including boundary treatments, enabling one pre-trained LNO to predict solutions on different domains.
For demonstration, LNO learns Navier-Stokes equations from randomly generated data samples, and then the pre-trained LNO is used as an explicit numerical time-marching scheme.
arXiv Detail & Related papers (2022-03-11T06:41:51Z) - Training Feedback Spiking Neural Networks by Implicit Differentiation on
the Equilibrium State [66.2457134675891]
Spiking neural networks (SNNs) are brain-inspired models that enable energy-efficient implementation on neuromorphic hardware.
Most existing methods imitate the backpropagation framework and feedforward architectures for artificial neural networks.
We propose a novel training method that does not rely on the exact reverse of the forward computation.
arXiv Detail & Related papers (2021-09-29T07:46:54Z) - Cell-average based neural network method for hyperbolic and parabolic
partial differential equations [0.0]
Motivated by finite volume scheme, a cell-average based neural network method is proposed.
The cell-average based neural network method can sharply evolve contact discontinuity with almost zero numerical diffusion introduced.
Shock and rarefaction waves are well captured for nonlinear hyperbolic conservation laws.
arXiv Detail & Related papers (2021-07-02T03:29:45Z) - Provably Efficient Neural Estimation of Structural Equation Model: An
Adversarial Approach [144.21892195917758]
We study estimation in a class of generalized Structural equation models (SEMs)
We formulate the linear operator equation as a min-max game, where both players are parameterized by neural networks (NNs), and learn the parameters of these neural networks using a gradient descent.
For the first time we provide a tractable estimation procedure for SEMs based on NNs with provable convergence and without the need for sample splitting.
arXiv Detail & Related papers (2020-07-02T17:55:47Z) - Multipole Graph Neural Operator for Parametric Partial Differential
Equations [57.90284928158383]
One of the main challenges in using deep learning-based methods for simulating physical systems is formulating physics-based data.
We propose a novel multi-level graph neural network framework that captures interaction at all ranges with only linear complexity.
Experiments confirm our multi-graph network learns discretization-invariant solution operators to PDEs and can be evaluated in linear time.
arXiv Detail & Related papers (2020-06-16T21:56:22Z) - Communication-Efficient Distributed Stochastic AUC Maximization with
Deep Neural Networks [50.42141893913188]
We study a distributed variable for large-scale AUC for a neural network as with a deep neural network.
Our model requires a much less number of communication rounds and still a number of communication rounds in theory.
Our experiments on several datasets show the effectiveness of our theory and also confirm our theory.
arXiv Detail & Related papers (2020-05-05T18:08:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.