The Deep Arbitrary Polynomial Chaos Neural Network or how Deep
Artificial Neural Networks could benefit from Data-Driven Homogeneous Chaos
Theory
- URL: http://arxiv.org/abs/2306.14753v1
- Date: Mon, 26 Jun 2023 15:09:14 GMT
- Title: The Deep Arbitrary Polynomial Chaos Neural Network or how Deep
Artificial Neural Networks could benefit from Data-Driven Homogeneous Chaos
Theory
- Authors: Sergey Oladyshkin, Timothy Praditia, Ilja Kr\"oker, Farid Mohammadi,
Wolfgang Nowak, Sebastian Otte
- Abstract summary: Approaches based on Deep Artificial Networks (DANN) are very popular in our days.
For a majority of deep learning approaches based on DANNs, the kernel structure of neural signal processing remains the same.
To tackle the challenge, we suggest to employ the data-driven generalization of PCE theory known as arbitrary chaos.
- Score: 0.44040106718326594
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Artificial Intelligence and Machine learning have been widely used in various
fields of mathematical computing, physical modeling, computational science,
communication science, and stochastic analysis. Approaches based on Deep
Artificial Neural Networks (DANN) are very popular in our days. Depending on
the learning task, the exact form of DANNs is determined via their multi-layer
architecture, activation functions and the so-called loss function. However,
for a majority of deep learning approaches based on DANNs, the kernel structure
of neural signal processing remains the same, where the node response is
encoded as a linear superposition of neural activity, while the non-linearity
is triggered by the activation functions. In the current paper, we suggest to
analyze the neural signal processing in DANNs from the point of view of
homogeneous chaos theory as known from polynomial chaos expansion (PCE). From
the PCE perspective, the (linear) response on each node of a DANN could be seen
as a $1^{st}$ degree multi-variate polynomial of single neurons from the
previous layer, i.e. linear weighted sum of monomials. From this point of view,
the conventional DANN structure relies implicitly (but erroneously) on a
Gaussian distribution of neural signals. Additionally, this view revels that by
design DANNs do not necessarily fulfill any orthogonality or orthonormality
condition for a majority of data-driven applications. Therefore, the prevailing
handling of neural signals in DANNs could lead to redundant representation as
any neural signal could contain some partial information from other neural
signals. To tackle that challenge, we suggest to employ the data-driven
generalization of PCE theory known as arbitrary polynomial chaos (aPC) to
construct a corresponding multi-variate orthonormal representations on each
node of a DANN to obtain Deep arbitrary polynomial chaos neural networks.
Related papers
- A frugal Spiking Neural Network for unsupervised classification of continuous multivariate temporal data [0.0]
Spiking Neural Networks (SNNs) are neuromorphic and use more biologically plausible neurons with evolving membrane potentials.
We introduce here a frugal single-layer SNN designed for fully unsupervised identification and classification of multivariate temporal patterns in continuous data.
arXiv Detail & Related papers (2024-08-08T08:15:51Z) - Novel Kernel Models and Exact Representor Theory for Neural Networks Beyond the Over-Parameterized Regime [52.00917519626559]
This paper presents two models of neural-networks and their training applicable to neural networks of arbitrary width, depth and topology.
We also present an exact novel representor theory for layer-wise neural network training with unregularized gradient descent in terms of a local-extrinsic neural kernel (LeNK)
This representor theory gives insight into the role of higher-order statistics in neural network training and the effect of kernel evolution in neural-network kernel models.
arXiv Detail & Related papers (2024-05-24T06:30:36Z) - Efficient, probabilistic analysis of combinatorial neural codes [0.0]
neural networks encode inputs in the form of combinations of individual neurons' activities.
These neural codes present a computational challenge due to their high dimensionality and often large volumes of data.
We apply methods previously applied to small examples and apply them to large neural codes generated by experiments.
arXiv Detail & Related papers (2022-10-19T11:58:26Z) - Extrapolation and Spectral Bias of Neural Nets with Hadamard Product: a
Polynomial Net Study [55.12108376616355]
The study on NTK has been devoted to typical neural network architectures, but is incomplete for neural networks with Hadamard products (NNs-Hp)
In this work, we derive the finite-width-K formulation for a special class of NNs-Hp, i.e., neural networks.
We prove their equivalence to the kernel regression predictor with the associated NTK, which expands the application scope of NTK.
arXiv Detail & Related papers (2022-09-16T06:36:06Z) - Interpretable Polynomial Neural Ordinary Differential Equations [3.04585143845864]
We introduce the neural ODE, which is a deep neural network inside of the neural ODE framework.
We demonstrate the capability of neural ODEs to predict outside of the training region, as well as perform direct symbolic regression without additional tools such as SINDy.
arXiv Detail & Related papers (2022-08-09T23:23:37Z) - Nonlocal Kernel Network (NKN): a Stable and Resolution-Independent Deep
Neural Network [23.465930256410722]
Nonlocal kernel network (NKN) is resolution independent, characterized by deep neural networks.
NKN is capable of handling a variety of tasks such as learning governing equations and classifying images.
arXiv Detail & Related papers (2022-01-06T19:19:35Z) - The Separation Capacity of Random Neural Networks [78.25060223808936]
We show that a sufficiently large two-layer ReLU-network with standard Gaussian weights and uniformly distributed biases can solve this problem with high probability.
We quantify the relevant structure of the data in terms of a novel notion of mutual complexity.
arXiv Detail & Related papers (2021-07-31T10:25:26Z) - Local Extreme Learning Machines and Domain Decomposition for Solving
Linear and Nonlinear Partial Differential Equations [0.0]
We present a neural network-based method for solving linear and nonlinear partial differential equations.
The method combines the ideas of extreme learning machines (ELM), domain decomposition and local neural networks.
We compare the current method with the deep Galerkin method (DGM) and the physics-informed neural network (PINN) in terms of the accuracy and computational cost.
arXiv Detail & Related papers (2020-12-04T23:19:39Z) - Deep Polynomial Neural Networks [77.70761658507507]
$Pi$Nets are a new class of function approximators based on expansions.
$Pi$Nets produce state-the-art results in three challenging tasks, i.e. image generation, face verification and 3D mesh representation learning.
arXiv Detail & Related papers (2020-06-20T16:23:32Z) - Multipole Graph Neural Operator for Parametric Partial Differential
Equations [57.90284928158383]
One of the main challenges in using deep learning-based methods for simulating physical systems is formulating physics-based data.
We propose a novel multi-level graph neural network framework that captures interaction at all ranges with only linear complexity.
Experiments confirm our multi-graph network learns discretization-invariant solution operators to PDEs and can be evaluated in linear time.
arXiv Detail & Related papers (2020-06-16T21:56:22Z) - Non-linear Neurons with Human-like Apical Dendrite Activations [81.18416067005538]
We show that a standard neuron followed by our novel apical dendrite activation (ADA) can learn the XOR logical function with 100% accuracy.
We conduct experiments on six benchmark data sets from computer vision, signal processing and natural language processing.
arXiv Detail & Related papers (2020-02-02T21:09:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.