Identifying Learning Rules From Neural Network Observables
- URL: http://arxiv.org/abs/2010.11765v2
- Date: Tue, 8 Dec 2020 18:48:02 GMT
- Title: Identifying Learning Rules From Neural Network Observables
- Authors: Aran Nayebi, Sanjana Srivastava, Surya Ganguli, Daniel L.K. Yamins
- Abstract summary: We show that different classes of learning rules can be separated solely on the basis of aggregate statistics of the weights, activations, or instantaneous layer-wise activity changes.
Our results suggest that activation patterns, available from electrophysiological recordings of post-synaptic activities, may provide a good basis on which to identify learning rules.
- Score: 26.96375335939315
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The brain modifies its synaptic strengths during learning in order to better
adapt to its environment. However, the underlying plasticity rules that govern
learning are unknown. Many proposals have been suggested, including Hebbian
mechanisms, explicit error backpropagation, and a variety of alternatives. It
is an open question as to what specific experimental measurements would need to
be made to determine whether any given learning rule is operative in a real
biological system. In this work, we take a "virtual experimental" approach to
this problem. Simulating idealized neuroscience experiments with artificial
neural networks, we generate a large-scale dataset of learning trajectories of
aggregate statistics measured in a variety of neural network architectures,
loss functions, learning rule hyperparameters, and parameter initializations.
We then take a discriminative approach, training linear and simple non-linear
classifiers to identify learning rules from features based on these
observables. We show that different classes of learning rules can be separated
solely on the basis of aggregate statistics of the weights, activations, or
instantaneous layer-wise activity changes, and that these results generalize to
limited access to the trajectory and held-out architectures and learning
curricula. We identify the statistics of each observable that are most relevant
for rule identification, finding that statistics from network activities across
training are more robust to unit undersampling and measurement noise than those
obtained from the synaptic strengths. Our results suggest that activation
patterns, available from electrophysiological recordings of post-synaptic
activities on the order of several hundred units, frequently measured at wider
intervals over the course of learning, may provide a good basis on which to
identify learning rules.
Related papers
- Provably Neural Active Learning Succeeds via Prioritizing Perplexing Samples [53.95282502030541]
Neural Network-based active learning (NAL) is a cost-effective data selection technique that utilizes neural networks to select and train on a small subset of samples.
We try to move one step forward by offering a unified explanation for the success of both query criteria-based NAL from a feature learning view.
arXiv Detail & Related papers (2024-06-06T10:38:01Z) - How connectivity structure shapes rich and lazy learning in neural
circuits [14.236853424595333]
We investigate how the structure of the initial weights -- in particular their effective rank -- influences the network learning regime.
Our research highlights the pivotal role of initial weight structures in shaping learning regimes.
arXiv Detail & Related papers (2023-10-12T17:08:45Z) - Understanding Activation Patterns in Artificial Neural Networks by
Exploring Stochastic Processes [0.0]
We propose utilizing the framework of processes, which has been underutilized thus far.
We focus solely on activation frequency, leveraging neuroscience techniques used for real neuron spike trains.
We derive parameters describing activation patterns in each network, revealing consistent differences across architectures and training sets.
arXiv Detail & Related papers (2023-08-01T22:12:30Z) - Measures of Information Reflect Memorization Patterns [53.71420125627608]
We show that the diversity in the activation patterns of different neurons is reflective of model generalization and memorization.
Importantly, we discover that information organization points to the two forms of memorization, even for neural activations computed on unlabelled in-distribution examples.
arXiv Detail & Related papers (2022-10-17T20:15:24Z) - Distinguishing Learning Rules with Brain Machine Interfaces [0.0]
We consider biologically plausible supervised- and reinforcement-learning rules.
We derive a metric to distinguish between learning rules by observing changes in the network activity during learning.
arXiv Detail & Related papers (2022-06-27T16:58:30Z) - Data-driven emergence of convolutional structure in neural networks [83.4920717252233]
We show how fully-connected neural networks solving a discrimination task can learn a convolutional structure directly from their inputs.
By carefully designing data models, we show that the emergence of this pattern is triggered by the non-Gaussian, higher-order local structure of the inputs.
arXiv Detail & Related papers (2022-02-01T17:11:13Z) - Neuronal Learning Analysis using Cycle-Consistent Adversarial Networks [4.874780144224057]
We use a variant of deep generative models called - CycleGAN, to learn the unknown mapping between pre- and post-learning neural activities.
We develop an end-to-end pipeline to preprocess, train and evaluate calcium fluorescence signals, and a procedure to interpret the resulting deep learning models.
arXiv Detail & Related papers (2021-11-25T13:24:19Z) - Dynamic Neural Diversification: Path to Computationally Sustainable
Neural Networks [68.8204255655161]
Small neural networks with a constrained number of trainable parameters, can be suitable resource-efficient candidates for many simple tasks.
We explore the diversity of the neurons within the hidden layer during the learning process.
We analyze how the diversity of the neurons affects predictions of the model.
arXiv Detail & Related papers (2021-09-20T15:12:16Z) - Gone Fishing: Neural Active Learning with Fisher Embeddings [55.08537975896764]
There is an increasing need for active learning algorithms that are compatible with deep neural networks.
This article introduces BAIT, a practical representation of tractable, and high-performing active learning algorithm for neural networks.
arXiv Detail & Related papers (2021-06-17T17:26:31Z) - On the Robustness of Active Learning [0.7340017786387767]
Active Learning is concerned with how to identify the most useful samples for a Machine Learning algorithm to be trained with.
We find that it is often applied with not enough care and domain knowledge.
We propose the new "Sum of Squared Logits" method based on the Simpson diversity index and investigate the effect of using the confusion matrix for balancing in sample selection.
arXiv Detail & Related papers (2020-06-18T09:07:23Z) - Towards Efficient Processing and Learning with Spikes: New Approaches
for Multi-Spike Learning [59.249322621035056]
We propose two new multi-spike learning rules which demonstrate better performance over other baselines on various tasks.
In the feature detection task, we re-examine the ability of unsupervised STDP with its limitations being presented.
Our proposed learning rules can reliably solve the task over a wide range of conditions without specific constraints being applied.
arXiv Detail & Related papers (2020-05-02T06:41:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.