The Spectral Bias of Polynomial Neural Networks
- URL: http://arxiv.org/abs/2202.13473v1
- Date: Sun, 27 Feb 2022 23:12:43 GMT
- Title: The Spectral Bias of Polynomial Neural Networks
- Authors: Moulik Choraria, Leello Tadesse Dadi, Grigorios Chrysos, Julien
Mairal, Volkan Cevher
- Abstract summary: Polynomial neural networks (PNNs) have been shown to be particularly effective at image generation and face recognition, where high-frequency information is critical.
Previous studies have revealed that neural networks demonstrate a $textitspectral bias$ towards low-frequency functions, which yields faster learning of low-frequency components during training.
Inspired by such studies, we conduct a spectral analysis of the Tangent Kernel (NTK) of PNNs.
We find that the $Pi$-Net family, i.e., a recently proposed parametrization of PNNs, speeds up the
- Score: 63.27903166253743
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Polynomial neural networks (PNNs) have been recently shown to be particularly
effective at image generation and face recognition, where high-frequency
information is critical. Previous studies have revealed that neural networks
demonstrate a $\textit{spectral bias}$ towards low-frequency functions, which
yields faster learning of low-frequency components during training. Inspired by
such studies, we conduct a spectral analysis of the Neural Tangent Kernel (NTK)
of PNNs. We find that the $\Pi$-Net family, i.e., a recently proposed
parametrization of PNNs, speeds up the learning of the higher frequencies. We
verify the theoretical bias through extensive experiments. We expect our
analysis to provide novel insights into designing architectures and learning
frameworks by incorporating multiplicative interactions via polynomials.
Related papers
- Addressing Spectral Bias of Deep Neural Networks by Multi-Grade Deep Learning [3.0468273116892752]
Deep neural networks (DNNs) exhibit a tendency to prioritize the learning of lower-frequency components of a function, struggling to capture its high-frequency features.
We propose to learn a function containing high-frequency components by composing several SNNs, each of which learns certain low-frequency information from the given data.
Our study reveals that MGDL excels at representing functions containing high-frequency information.
arXiv Detail & Related papers (2024-10-21T15:34:33Z) - How neural networks learn to classify chaotic time series [77.34726150561087]
We study the inner workings of neural networks trained to classify regular-versus-chaotic time series.
We find that the relation between input periodicity and activation periodicity is key for the performance of LKCNN models.
arXiv Detail & Related papers (2023-06-04T08:53:27Z) - Gradient Descent in Neural Networks as Sequential Learning in RKBS [63.011641517977644]
We construct an exact power-series representation of the neural network in a finite neighborhood of the initial weights.
We prove that, regardless of width, the training sequence produced by gradient descent can be exactly replicated by regularized sequential learning.
arXiv Detail & Related papers (2023-02-01T03:18:07Z) - Extrapolation and Spectral Bias of Neural Nets with Hadamard Product: a
Polynomial Net Study [55.12108376616355]
The study on NTK has been devoted to typical neural network architectures, but is incomplete for neural networks with Hadamard products (NNs-Hp)
In this work, we derive the finite-width-K formulation for a special class of NNs-Hp, i.e., neural networks.
We prove their equivalence to the kernel regression predictor with the associated NTK, which expands the application scope of NTK.
arXiv Detail & Related papers (2022-09-16T06:36:06Z) - On the Study of Sample Complexity for Polynomial Neural Networks [13.265045615849099]
Among various kinds of neural networks architectures, sample neural networks (PNNs) have been recently shown to be analyzable by spectrum analysis.
In this paper, we extend the analysis in previous literature to PNNs and obtain novel results on sample complexity of PNNs.
arXiv Detail & Related papers (2022-07-18T19:10:53Z) - Momentum Diminishes the Effect of Spectral Bias in Physics-Informed
Neural Networks [72.09574528342732]
Physics-informed neural network (PINN) algorithms have shown promising results in solving a wide range of problems involving partial differential equations (PDEs)
They often fail to converge to desirable solutions when the target function contains high-frequency features, due to a phenomenon known as spectral bias.
In the present work, we exploit neural tangent kernels (NTKs) to investigate the training dynamics of PINNs evolving under gradient descent with momentum (SGDM)
arXiv Detail & Related papers (2022-06-29T19:03:10Z) - Neural Tangent Kernel Analysis of Deep Narrow Neural Networks [11.623483126242478]
We present the first trainability guarantee of infinitely deep but narrow neural networks.
We then extend the analysis to an infinitely deep convolutional neural network (CNN) and perform brief experiments.
arXiv Detail & Related papers (2022-02-07T07:27:02Z) - Rectified Linear Postsynaptic Potential Function for Backpropagation in
Deep Spiking Neural Networks [55.0627904986664]
Spiking Neural Networks (SNNs) usetemporal spike patterns to represent and transmit information, which is not only biologically realistic but also suitable for ultra-low-power event-driven neuromorphic implementation.
This paper investigates the contribution of spike timing dynamics to information encoding, synaptic plasticity and decision making, providing a new perspective to design of future DeepSNNs and neuromorphic hardware systems.
arXiv Detail & Related papers (2020-03-26T11:13:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.