Spectral Neural Networks: Approximation Theory and Optimization
Landscape
- URL: http://arxiv.org/abs/2310.00729v1
- Date: Sun, 1 Oct 2023 17:03:47 GMT
- Title: Spectral Neural Networks: Approximation Theory and Optimization
Landscape
- Authors: Chenghui Li, Rishi Sonthalia, Nicolas Garcia Trillos
- Abstract summary: We present key theoretical aspects of Spectral Neural Network (SNN) training.
First, we present quantitative insights into the tradeoff between the number of neurons and the amount of spectral information a neural network learns.
- Score: 6.967392207053043
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: There is a large variety of machine learning methodologies that are based on
the extraction of spectral geometric information from data. However, the
implementations of many of these methods often depend on traditional
eigensolvers, which present limitations when applied in practical online big
data scenarios. To address some of these challenges, researchers have proposed
different strategies for training neural networks as alternatives to
traditional eigensolvers, with one such approach known as Spectral Neural
Network (SNN). In this paper, we investigate key theoretical aspects of SNN.
First, we present quantitative insights into the tradeoff between the number of
neurons and the amount of spectral geometric information a neural network
learns. Second, we initiate a theoretical exploration of the optimization
landscape of SNN's objective to shed light on the training dynamics of SNN.
Unlike typical studies of convergence to global solutions of NN training
dynamics, SNN presents an additional complexity due to its non-convex ambient
loss function.
Related papers
- Evolutionary Spiking Neural Networks: A Survey [21.27024959576232]
Spiking neural networks (SNNs) are gaining increasing attention as potential computationally efficient alternatives to traditional artificial neural networks(ANNs)
The unique information propagation mechanisms and the complexity of SNN neuron models pose challenges for adopting traditional methods developed for ANNs to SNNs.
Recently, a novel paradigm utilizing evolutionary computation methods has emerged to tackle these challenges.
arXiv Detail & Related papers (2024-06-18T12:27:55Z) - Direct Training High-Performance Deep Spiking Neural Networks: A Review of Theories and Methods [33.377770671553336]
Spiking neural networks (SNNs) offer a promising energy-efficient alternative to artificial neural networks (ANNs)
In this paper, we provide a new perspective to summarize the theories and methods for training deep SNNs with high performance.
arXiv Detail & Related papers (2024-05-06T09:58:54Z) - Deep Neural Networks via Complex Network Theory: a Perspective [3.1023851130450684]
Deep Neural Networks (DNNs) can be represented as graphs whose links and vertices iteratively process data and solve tasks sub-optimally. Complex Network Theory (CNT), merging statistical physics with graph theory, provides a method for interpreting neural networks by analysing their weights and neuron structures.
In this work, we extend the existing CNT metrics with measures that sample from the DNNs' training distribution, shifting from a purely topological analysis to one that connects with the interpretability of deep learning.
arXiv Detail & Related papers (2024-04-17T08:42:42Z) - Deeper or Wider: A Perspective from Optimal Generalization Error with Sobolev Loss [2.07180164747172]
We compare deeper neural networks (DeNNs) with a flexible number of layers and wider neural networks (WeNNs) with limited hidden layers.
We find that a higher number of parameters tends to favor WeNNs, while an increased number of sample points and greater regularity in the loss function lean towards the adoption of DeNNs.
arXiv Detail & Related papers (2024-01-31T20:10:10Z) - A Hybrid Neural Coding Approach for Pattern Recognition with Spiking
Neural Networks [53.31941519245432]
Brain-inspired spiking neural networks (SNNs) have demonstrated promising capabilities in solving pattern recognition tasks.
These SNNs are grounded on homogeneous neurons that utilize a uniform neural coding for information representation.
In this study, we argue that SNN architectures should be holistically designed to incorporate heterogeneous coding schemes.
arXiv Detail & Related papers (2023-05-26T02:52:12Z) - Extrapolation and Spectral Bias of Neural Nets with Hadamard Product: a
Polynomial Net Study [55.12108376616355]
The study on NTK has been devoted to typical neural network architectures, but is incomplete for neural networks with Hadamard products (NNs-Hp)
In this work, we derive the finite-width-K formulation for a special class of NNs-Hp, i.e., neural networks.
We prove their equivalence to the kernel regression predictor with the associated NTK, which expands the application scope of NTK.
arXiv Detail & Related papers (2022-09-16T06:36:06Z) - On the Study of Sample Complexity for Polynomial Neural Networks [13.265045615849099]
Among various kinds of neural networks architectures, sample neural networks (PNNs) have been recently shown to be analyzable by spectrum analysis.
In this paper, we extend the analysis in previous literature to PNNs and obtain novel results on sample complexity of PNNs.
arXiv Detail & Related papers (2022-07-18T19:10:53Z) - Knowledge Enhanced Neural Networks for relational domains [83.9217787335878]
We focus on a specific method, KENN, a Neural-Symbolic architecture that injects prior logical knowledge into a neural network.
In this paper, we propose an extension of KENN for relational data.
arXiv Detail & Related papers (2022-05-31T13:00:34Z) - The Spectral Bias of Polynomial Neural Networks [63.27903166253743]
Polynomial neural networks (PNNs) have been shown to be particularly effective at image generation and face recognition, where high-frequency information is critical.
Previous studies have revealed that neural networks demonstrate a $textitspectral bias$ towards low-frequency functions, which yields faster learning of low-frequency components during training.
Inspired by such studies, we conduct a spectral analysis of the Tangent Kernel (NTK) of PNNs.
We find that the $Pi$-Net family, i.e., a recently proposed parametrization of PNNs, speeds up the
arXiv Detail & Related papers (2022-02-27T23:12:43Z) - Modeling from Features: a Mean-field Framework for Over-parameterized
Deep Neural Networks [54.27962244835622]
This paper proposes a new mean-field framework for over- parameterized deep neural networks (DNNs)
In this framework, a DNN is represented by probability measures and functions over its features in the continuous limit.
We illustrate the framework via the standard DNN and the Residual Network (Res-Net) architectures.
arXiv Detail & Related papers (2020-07-03T01:37:16Z) - Rectified Linear Postsynaptic Potential Function for Backpropagation in
Deep Spiking Neural Networks [55.0627904986664]
Spiking Neural Networks (SNNs) usetemporal spike patterns to represent and transmit information, which is not only biologically realistic but also suitable for ultra-low-power event-driven neuromorphic implementation.
This paper investigates the contribution of spike timing dynamics to information encoding, synaptic plasticity and decision making, providing a new perspective to design of future DeepSNNs and neuromorphic hardware systems.
arXiv Detail & Related papers (2020-03-26T11:13:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.