Testing predictions of representation cost theory with CNNs
- URL: http://arxiv.org/abs/2210.01257v3
- Date: Tue, 26 Sep 2023 03:09:25 GMT
- Title: Testing predictions of representation cost theory with CNNs
- Authors: Charles Godfrey, Elise Bishoff, Myles Mckay, Davis Brown, Grayson
Jorgenson, Henry Kvinge and Eleanor Byler
- Abstract summary: We show that trained convolutional neural networks (CNNs) have different levels of sensitivity to signals of different frequency.
This is a consequence of the frequency distribution of natural images, which is known to have most of its power concentrated in low-to-mid frequencies.
- Score: 5.816527700115096
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: It is widely acknowledged that trained convolutional neural networks (CNNs)
have different levels of sensitivity to signals of different frequency. In
particular, a number of empirical studies have documented CNNs sensitivity to
low-frequency signals. In this work we show with theory and experiments that
this observed sensitivity is a consequence of the frequency distribution of
natural images, which is known to have most of its power concentrated in
low-to-mid frequencies. Our theoretical analysis relies on representations of
the layers of a CNN in frequency space, an idea that has previously been used
to accelerate computations and study implicit bias of network training
algorithms, but to the best of our knowledge has not been applied in the domain
of model robustness.
Related papers
- A rationale from frequency perspective for grokking in training neural network [7.264378254137811]
Grokking is the phenomenon where neural networks NNs initially fit the training data and later generalize to the test data during training.
In this paper, we empirically provide a frequency perspective to explain the emergence of this phenomenon in NNs.
arXiv Detail & Related papers (2024-05-24T06:57:23Z) - Understanding the dynamics of the frequency bias in neural networks [0.0]
Recent works have shown that traditional Neural Network (NN) architectures display a marked frequency bias in the learning process.
We develop a partial differential equation (PDE) that unravels the frequency dynamics of the error for a 2-layer NN.
We empirically show that the same principle extends to multi-layer NNs.
arXiv Detail & Related papers (2024-05-23T18:09:16Z) - Deep Neural Networks Tend To Extrapolate Predictably [51.303814412294514]
neural network predictions tend to be unpredictable and overconfident when faced with out-of-distribution (OOD) inputs.
We observe that neural network predictions often tend towards a constant value as input data becomes increasingly OOD.
We show how one can leverage our insights in practice to enable risk-sensitive decision-making in the presence of OOD inputs.
arXiv Detail & Related papers (2023-10-02T03:25:32Z) - How neural networks learn to classify chaotic time series [77.34726150561087]
We study the inner workings of neural networks trained to classify regular-versus-chaotic time series.
We find that the relation between input periodicity and activation periodicity is key for the performance of LKCNN models.
arXiv Detail & Related papers (2023-06-04T08:53:27Z) - A Scalable Walsh-Hadamard Regularizer to Overcome the Low-degree
Spectral Bias of Neural Networks [79.28094304325116]
Despite the capacity of neural nets to learn arbitrary functions, models trained through gradient descent often exhibit a bias towards simpler'' functions.
We show how this spectral bias towards low-degree frequencies can in fact hurt the neural network's generalization on real-world datasets.
We propose a new scalable functional regularization scheme that aids the neural network to learn higher degree frequencies.
arXiv Detail & Related papers (2023-05-16T20:06:01Z) - Frequency and Scale Perspectives of Feature Extraction [5.081561820537235]
We analyze the sensitivity of neural networks to frequencies and scales.
We find that neural networks have low- and medium-frequency biases but also prefer different frequency bands for different classes.
These observations lead to the hypothesis that neural networks must learn the ability to extract features at various scales and frequencies.
arXiv Detail & Related papers (2023-02-24T06:37:36Z) - Momentum Diminishes the Effect of Spectral Bias in Physics-Informed
Neural Networks [72.09574528342732]
Physics-informed neural network (PINN) algorithms have shown promising results in solving a wide range of problems involving partial differential equations (PDEs)
They often fail to converge to desirable solutions when the target function contains high-frequency features, due to a phenomenon known as spectral bias.
In the present work, we exploit neural tangent kernels (NTKs) to investigate the training dynamics of PINNs evolving under gradient descent with momentum (SGDM)
arXiv Detail & Related papers (2022-06-29T19:03:10Z) - The Spectral Bias of Polynomial Neural Networks [63.27903166253743]
Polynomial neural networks (PNNs) have been shown to be particularly effective at image generation and face recognition, where high-frequency information is critical.
Previous studies have revealed that neural networks demonstrate a $textitspectral bias$ towards low-frequency functions, which yields faster learning of low-frequency components during training.
Inspired by such studies, we conduct a spectral analysis of the Tangent Kernel (NTK) of PNNs.
We find that the $Pi$-Net family, i.e., a recently proposed parametrization of PNNs, speeds up the
arXiv Detail & Related papers (2022-02-27T23:12:43Z) - Deep Networks for Direction-of-Arrival Estimation in Low SNR [89.45026632977456]
We introduce a Convolutional Neural Network (CNN) that is trained from mutli-channel data of the true array manifold matrix.
We train a CNN in the low-SNR regime to predict DoAs across all SNRs.
Our robust solution can be applied in several fields, ranging from wireless array sensors to acoustic microphones or sonars.
arXiv Detail & Related papers (2020-11-17T12:52:18Z) - The use of Convolutional Neural Networks for signal-background
classification in Particle Physics experiments [0.4301924025274017]
We present an extensive convolutional neural architecture search, achieving high accuracy for signal/background discrimination for a HEP classification use-case.
We demonstrate among other things that we can achieve the same accuracy as complex ResNet architectures with CNNs with less parameters.
arXiv Detail & Related papers (2020-02-13T19:54:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.