Weakly-correlated synapses promote dimension reduction in deep neural
networks
- URL: http://arxiv.org/abs/2006.11569v1
- Date: Sat, 20 Jun 2020 13:11:37 GMT
- Title: Weakly-correlated synapses promote dimension reduction in deep neural
networks
- Authors: Jianwen Zhou, and Haiping Huang
- Abstract summary: How synaptic correlations affect neural correlations to produce disentangled hidden representations remains elusive.
We propose a model of dimension reduction, taking into account pairwise correlations among synapses.
Our theory determines the synaptic-correlation scaling form requiring only mathematical self-consistency.
- Score: 1.7532045941271799
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: By controlling synaptic and neural correlations, deep learning has achieved
empirical successes in improving classification performances. How synaptic
correlations affect neural correlations to produce disentangled hidden
representations remains elusive. Here we propose a simplified model of
dimension reduction, taking into account pairwise correlations among synapses,
to reveal the mechanism underlying how the synaptic correlations affect
dimension reduction. Our theory determines the synaptic-correlation scaling
form requiring only mathematical self-consistency, for both binary and
continuous synapses. The theory also predicts that weakly-correlated synapses
encourage dimension reduction compared to their orthogonal counterparts. In
addition, these synapses slow down the decorrelation process along the network
depth. These two computational roles are explained by the proposed mean-field
equation. The theoretical predictions are in excellent agreement with numerical
simulations, and the key features are also captured by a deep learning with
Hebbian rules.
Related papers
- Artificial Kuramoto Oscillatory Neurons [65.16453738828672]
We introduce Artificial Kuramotoy Neurons (AKOrN) as a dynamical alternative to threshold units.
We show that this idea provides performance improvements across a wide spectrum of tasks.
We believe that these empirical results show the importance of our assumptions at the most basic neuronal level of neural representation.
arXiv Detail & Related papers (2024-10-17T17:47:54Z) - Connecting Concept Convexity and Human-Machine Alignment in Deep Neural Networks [3.001674556825579]
Understanding how neural networks align with human cognitive processes is a crucial step toward developing more interpretable and reliable AI systems.
We identify a correlation between these two dimensions that reflect the similarity relations humans in cognitive tasks.
This presents a first step toward understanding the relationship convexity between human-machine alignment.
arXiv Detail & Related papers (2024-09-10T09:32:16Z) - Exploring neural oscillations during speech perception via surrogate gradient spiking neural networks [59.38765771221084]
We present a physiologically inspired speech recognition architecture compatible and scalable with deep learning frameworks.
We show end-to-end gradient descent training leads to the emergence of neural oscillations in the central spiking neural network.
Our findings highlight the crucial inhibitory role of feedback mechanisms, such as spike frequency adaptation and recurrent connections, in regulating and synchronising neural activity to improve recognition performance.
arXiv Detail & Related papers (2024-04-22T09:40:07Z) - A simple connection from loss flatness to compressed representations in neural networks [3.5502600490147196]
We show that in the final phase of learning in deep neural networks, the compression of the manifold of neural representations correlates with the flatness of the loss around the minima explored by SGD.
Our work builds upon the linear stability insight by Ma and Ying, deriving inequalities between various compression metrics and quantities involving sharpness.
arXiv Detail & Related papers (2023-10-03T03:36:29Z) - Hybrid Predictive Coding: Inferring, Fast and Slow [62.997667081978825]
We propose a hybrid predictive coding network that combines both iterative and amortized inference in a principled manner.
We demonstrate that our model is inherently sensitive to its uncertainty and adaptively balances balances to obtain accurate beliefs using minimum computational expense.
arXiv Detail & Related papers (2022-04-05T12:52:45Z) - Phenomenology of Double Descent in Finite-Width Neural Networks [29.119232922018732]
Double descent delineates the behaviour of models depending on the regime they belong to.
We use influence functions to derive suitable expressions of the population loss and its lower bound.
Building on our analysis, we investigate how the loss function affects double descent.
arXiv Detail & Related papers (2022-03-14T17:39:49Z) - Neuronal Correlation: a Central Concept in Neural Network [22.764342635264452]
We show that neuronal correlation can be efficiently estimated via weight matrix.
We show that neuronal correlation significantly impacts on the accuracy of entropy estimation in high-dimensional hidden spaces.
arXiv Detail & Related papers (2022-01-22T15:01:50Z) - The Interplay Between Implicit Bias and Benign Overfitting in Two-Layer
Linear Networks [51.1848572349154]
neural network models that perfectly fit noisy data can generalize well to unseen test data.
We consider interpolating two-layer linear neural networks trained with gradient flow on the squared loss and derive bounds on the excess risk.
arXiv Detail & Related papers (2021-08-25T22:01:01Z) - The distribution of inhibitory neurons in the C. elegans connectome
facilitates self-optimization of coordinated neural activity [78.15296214629433]
The nervous system of the nematode Caenorhabditis elegans exhibits remarkable complexity despite the worm's small size.
A general challenge is to better understand the relationship between neural organization and neural activity at the system level.
We implemented an abstract simulation model of the C. elegans connectome that approximates the neurotransmitter identity of each neuron.
arXiv Detail & Related papers (2020-10-28T23:11:37Z) - Optimal Learning with Excitatory and Inhibitory synapses [91.3755431537592]
I study the problem of storing associations between analog signals in the presence of correlations.
I characterize the typical learning performance in terms of the power spectrum of random input and output processes.
arXiv Detail & Related papers (2020-05-25T18:25:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.