Deep Neural Networks and Finite Elements of Any Order on Arbitrary
Dimensions
- URL: http://arxiv.org/abs/2312.14276v3
- Date: Thu, 11 Jan 2024 21:01:25 GMT
- Title: Deep Neural Networks and Finite Elements of Any Order on Arbitrary
Dimensions
- Authors: Juncai He, Jinchao Xu
- Abstract summary: Deep neural networks employing ReLU and ReLU$2$ activation functions can effectively represent Lagrange finite element functions of any order on various simplicial meshes in arbitrary dimensions.
Our findings present the first demonstration of how deep neural networks can systematically generate general continuous piecewise functions on both specific or arbitrary simplicial meshes.
- Score: 2.7195102129095003
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this study, we establish that deep neural networks employing ReLU and
ReLU$^2$ activation functions can effectively represent Lagrange finite element
functions of any order on various simplicial meshes in arbitrary dimensions. We
introduce two novel formulations for globally expressing the basis functions of
Lagrange elements, tailored for both specific and arbitrary meshes. These
formulations are based on a geometric decomposition of the elements,
incorporating several insightful and essential properties of high-dimensional
simplicial meshes, barycentric coordinate functions, and global basis functions
of linear elements. This representation theory facilitates a natural
approximation result for such deep neural networks. Our findings present the
first demonstration of how deep neural networks can systematically generate
general continuous piecewise polynomial functions on both specific or arbitrary
simplicial meshes.
Related papers
- Shallow ReLU neural networks and finite elements [1.3597551064547502]
We show that piecewise linear functions on a convex polytope mesh can be represented by two-hidden-layer ReLU neural networks in a weak sense.
The numbers of neurons of the two hidden layers required to weakly represent are accurately given based on the numbers of polytopes and hyperplanes involved in this mesh.
arXiv Detail & Related papers (2024-03-09T06:12:06Z) - Structure of universal formulas [13.794391803767617]
We introduce a hierarchy of classes connecting the global approximability property to the weaker property of infinite VC dimension.
We show that fixed-size neural networks with not more than one layer of neurons having activations cannot approximate functions on arbitrary finite sets.
We give examples of functional families, including two-hidden-layer neural networks, that approximate functions on arbitrary finite sets, but fail to do that on the whole domain of definition.
arXiv Detail & Related papers (2023-11-07T11:50:25Z) - Exploring Linear Feature Disentanglement For Neural Networks [63.20827189693117]
Non-linear activation functions, e.g., Sigmoid, ReLU, and Tanh, have achieved great success in neural networks (NNs)
Due to the complex non-linear characteristic of samples, the objective of those activation functions is to project samples from their original feature space to a linear separable feature space.
This phenomenon ignites our interest in exploring whether all features need to be transformed by all non-linear functions in current typical NNs.
arXiv Detail & Related papers (2022-03-22T13:09:17Z) - Dist2Cycle: A Simplicial Neural Network for Homology Localization [66.15805004725809]
Simplicial complexes can be viewed as high dimensional generalizations of graphs that explicitly encode multi-way ordered relations.
We propose a graph convolutional model for learning functions parametrized by the $k$-homological features of simplicial complexes.
arXiv Detail & Related papers (2021-10-28T14:59:41Z) - Geometry of Linear Convolutional Networks [7.990816079551592]
We study the family of functions represented by a linear convolutional neural network (LCN)
We study the optimization of an objective function over an LCN, analyzing critical points in function space and in gradient space.
Overall, our theory predicts that the optimized parameters of an LCN will often correspond to repeated filters across layers.
arXiv Detail & Related papers (2021-08-03T14:42:18Z) - Going Beyond Linear RL: Sample Efficient Neural Function Approximation [76.57464214864756]
We study function approximation with two-layer neural networks.
Our results significantly improve upon what can be attained with linear (or eluder dimension) methods.
arXiv Detail & Related papers (2021-07-14T03:03:56Z) - Principled Simplicial Neural Networks for Trajectory Prediction [17.016397531234393]
We consider the construction of neural network architectures for data on simplicial complexes.
Based on these properties, we propose a simple convolutional architecture for the problem of trajectory prediction.
We show that it obeys all three of these properties when an odd, nonlinear activation function is used.
arXiv Detail & Related papers (2021-02-19T17:37:43Z) - On the Number of Linear Functions Composing Deep Neural Network: Towards
a Refined Definition of Neural Networks Complexity [6.252236971703546]
We introduce an equivalence relation among the linear functions composing a piecewise linear function and then count those linear functions relative to that equivalence relation.
Our new complexity measure can clearly distinguish between the two models, is consistent with the classical measure, and increases exponentially with depth.
arXiv Detail & Related papers (2020-10-23T01:46:12Z) - UNIPoint: Universally Approximating Point Processes Intensities [125.08205865536577]
We provide a proof that a class of learnable functions can universally approximate any valid intensity function.
We implement UNIPoint, a novel neural point process model, using recurrent neural networks to parameterise sums of basis function upon each event.
arXiv Detail & Related papers (2020-07-28T09:31:56Z) - Multipole Graph Neural Operator for Parametric Partial Differential
Equations [57.90284928158383]
One of the main challenges in using deep learning-based methods for simulating physical systems is formulating physics-based data.
We propose a novel multi-level graph neural network framework that captures interaction at all ranges with only linear complexity.
Experiments confirm our multi-graph network learns discretization-invariant solution operators to PDEs and can be evaluated in linear time.
arXiv Detail & Related papers (2020-06-16T21:56:22Z) - Invariant Feature Coding using Tensor Product Representation [75.62232699377877]
We prove that the group-invariant feature vector contains sufficient discriminative information when learning a linear classifier.
A novel feature model that explicitly consider group action is proposed for principal component analysis and k-means clustering.
arXiv Detail & Related papers (2019-06-05T07:15:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.