Nonlocal Kernel Network (NKN): a Stable and Resolution-Independent Deep
Neural Network
- URL: http://arxiv.org/abs/2201.02217v1
- Date: Thu, 6 Jan 2022 19:19:35 GMT
- Title: Nonlocal Kernel Network (NKN): a Stable and Resolution-Independent Deep
Neural Network
- Authors: Huaiqian You, Yue Yu, Marta D'Elia, Tian Gao, Stewart Silling
- Abstract summary: Nonlocal kernel network (NKN) is resolution independent, characterized by deep neural networks.
NKN is capable of handling a variety of tasks such as learning governing equations and classifying images.
- Score: 23.465930256410722
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Neural operators have recently become popular tools for designing solution
maps between function spaces in the form of neural networks. Differently from
classical scientific machine learning approaches that learn parameters of a
known partial differential equation (PDE) for a single instance of the input
parameters at a fixed resolution, neural operators approximate the solution map
of a family of PDEs. Despite their success, the uses of neural operators are so
far restricted to relatively shallow neural networks and confined to learning
hidden governing laws. In this work, we propose a novel nonlocal neural
operator, which we refer to as nonlocal kernel network (NKN), that is
resolution independent, characterized by deep neural networks, and capable of
handling a variety of tasks such as learning governing equations and
classifying images. Our NKN stems from the interpretation of the neural network
as a discrete nonlocal diffusion reaction equation that, in the limit of
infinite layers, is equivalent to a parabolic nonlocal equation, whose
stability is analyzed via nonlocal vector calculus. The resemblance with
integral forms of neural operators allows NKNs to capture long-range
dependencies in the feature space, while the continuous treatment of
node-to-node interactions makes NKNs resolution independent. The resemblance
with neural ODEs, reinterpreted in a nonlocal sense, and the stable network
dynamics between layers allow for generalization of NKN's optimal parameters
from shallow to deep networks. This fact enables the use of shallow-to-deep
initialization techniques. Our tests show that NKNs outperform baseline methods
in both learning governing equations and image classification tasks and
generalize well to different resolutions and depths.
Related papers
- HyResPINNs: Adaptive Hybrid Residual Networks for Learning Optimal Combinations of Neural and RBF Components for Physics-Informed Modeling [22.689531776611084]
We present a new class of PINNs called HyResPINNs, which augment traditional PINNs with adaptive hybrid residual blocks.
A key feature of our method is the inclusion of adaptive combination parameters within each residual block.
We show that HyResPINNs are more robust to training point locations and neural network architectures than traditional PINNs.
arXiv Detail & Related papers (2024-10-04T16:21:14Z) - Novel Kernel Models and Exact Representor Theory for Neural Networks Beyond the Over-Parameterized Regime [52.00917519626559]
This paper presents two models of neural-networks and their training applicable to neural networks of arbitrary width, depth and topology.
We also present an exact novel representor theory for layer-wise neural network training with unregularized gradient descent in terms of a local-extrinsic neural kernel (LeNK)
This representor theory gives insight into the role of higher-order statistics in neural network training and the effect of kernel evolution in neural-network kernel models.
arXiv Detail & Related papers (2024-05-24T06:30:36Z) - Deeper or Wider: A Perspective from Optimal Generalization Error with Sobolev Loss [2.07180164747172]
We compare deeper neural networks (DeNNs) with a flexible number of layers and wider neural networks (WeNNs) with limited hidden layers.
We find that a higher number of parameters tends to favor WeNNs, while an increased number of sample points and greater regularity in the loss function lean towards the adoption of DeNNs.
arXiv Detail & Related papers (2024-01-31T20:10:10Z) - Extrapolation and Spectral Bias of Neural Nets with Hadamard Product: a
Polynomial Net Study [55.12108376616355]
The study on NTK has been devoted to typical neural network architectures, but is incomplete for neural networks with Hadamard products (NNs-Hp)
In this work, we derive the finite-width-K formulation for a special class of NNs-Hp, i.e., neural networks.
We prove their equivalence to the kernel regression predictor with the associated NTK, which expands the application scope of NTK.
arXiv Detail & Related papers (2022-09-16T06:36:06Z) - Neural Operator: Learning Maps Between Function Spaces [75.93843876663128]
We propose a generalization of neural networks to learn operators, termed neural operators, that map between infinite dimensional function spaces.
We prove a universal approximation theorem for our proposed neural operator, showing that it can approximate any given nonlinear continuous operator.
An important application for neural operators is learning surrogate maps for the solution operators of partial differential equations.
arXiv Detail & Related papers (2021-08-19T03:56:49Z) - Deep Kronecker neural networks: A general framework for neural networks
with adaptive activation functions [4.932130498861987]
We propose a new type of neural networks, Kronecker neural networks (KNNs), that form a general framework for neural networks with adaptive activation functions.
Under suitable conditions, KNNs induce a faster decay of the loss than that by the feed-forward networks.
arXiv Detail & Related papers (2021-05-20T04:54:57Z) - Local Extreme Learning Machines and Domain Decomposition for Solving
Linear and Nonlinear Partial Differential Equations [0.0]
We present a neural network-based method for solving linear and nonlinear partial differential equations.
The method combines the ideas of extreme learning machines (ELM), domain decomposition and local neural networks.
We compare the current method with the deep Galerkin method (DGM) and the physics-informed neural network (PINN) in terms of the accuracy and computational cost.
arXiv Detail & Related papers (2020-12-04T23:19:39Z) - Exploiting Heterogeneity in Operational Neural Networks by Synaptic
Plasticity [87.32169414230822]
Recently proposed network model, Operational Neural Networks (ONNs), can generalize the conventional Convolutional Neural Networks (CNNs)
In this study the focus is drawn on searching the best-possible operator set(s) for the hidden neurons of the network based on the Synaptic Plasticity paradigm that poses the essential learning theory in biological neurons.
Experimental results over highly challenging problems demonstrate that the elite ONNs even with few neurons and layers can achieve a superior learning performance than GIS-based ONNs.
arXiv Detail & Related papers (2020-08-21T19:03:23Z) - Modeling from Features: a Mean-field Framework for Over-parameterized
Deep Neural Networks [54.27962244835622]
This paper proposes a new mean-field framework for over- parameterized deep neural networks (DNNs)
In this framework, a DNN is represented by probability measures and functions over its features in the continuous limit.
We illustrate the framework via the standard DNN and the Residual Network (Res-Net) architectures.
arXiv Detail & Related papers (2020-07-03T01:37:16Z) - Self-Organized Operational Neural Networks with Generative Neurons [87.32169414230822]
ONNs are heterogenous networks with a generalized neuron model that can encapsulate any set of non-linear operators.
We propose Self-organized ONNs (Self-ONNs) with generative neurons that have the ability to adapt (optimize) the nodal operator of each connection.
arXiv Detail & Related papers (2020-04-24T14:37:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.