Statistical Mechanics of Learning via Reverberation in Bidirectional
Associative Memories
- URL: http://arxiv.org/abs/2307.08365v1
- Date: Mon, 17 Jul 2023 10:04:04 GMT
- Title: Statistical Mechanics of Learning via Reverberation in Bidirectional
Associative Memories
- Authors: Martino Salomone Centonze, Ido Kanter, Adriano Barra
- Abstract summary: We study bi-directional associative neural networks that are exposed to noisy examples of random archetypes.
In this setting, learning is heteroassociative -- involving couples of patterns -- and it is achieved by reverberating the information depicted from the examples.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We study bi-directional associative neural networks that, exposed to noisy
examples of an extensive number of random archetypes, learn the latter (with or
without the presence of a teacher) when the supplied information is enough: in
this setting, learning is heteroassociative -- involving couples of patterns --
and it is achieved by reverberating the information depicted from the examples
through the layers of the network. By adapting Guerra's interpolation
technique, we provide a full statistical mechanical picture of supervised and
unsupervised learning processes (at the replica symmetric level of description)
obtaining analytically phase diagrams, thresholds for learning, a picture of
the ground-state in plain agreement with Monte Carlo simulations and
signal-to-noise outcomes. In the large dataset limit, the Kosko storage
prescription as well as its statistical mechanical picture provided by Kurchan,
Peliti, and Saber in the eighties is fully recovered. Computational advantages
in dealing with information reverberation, rather than storage, are discussed
for natural test cases. In particular, we show how this network admits an
integral representation in terms of two coupled restricted Boltzmann machines,
whose hidden layers are entirely built of by grand-mother neurons, to prove
that by coupling solely these grand-mother neurons we can correlate the
patterns they are related to: it is thus possible to recover Pavlov's Classical
Conditioning by adding just one synapse among the correct grand-mother neurons
(hence saving an extensive number of these links for further information
storage w.r.t. the classical autoassociative setting).
Related papers
- Demolition and Reinforcement of Memories in Spin-Glass-like Neural
Networks [0.0]
The aim of this thesis is to understand the effectiveness of Unlearning in both associative memory models and generative models.
The selection of structured data enables an associative memory model to retrieve concepts as attractors of a neural dynamics with considerable basins of attraction.
A novel regularization technique for Boltzmann Machines is presented, proving to outperform previously developed methods in learning hidden probability distributions from data-sets.
arXiv Detail & Related papers (2024-03-04T23:12:42Z) - Hebbian Learning from First Principles [0.0]
We postulating the expression of its Hamiltonian for supervised and unsupervised protocols.
We show how Lagrangian constraints within entropy extremization force network's outcomes on neural correlations.
Remarks on the exponential Hopfield model (as the limit of dense networks with diverging density) and semi-supervised protocols are also provided.
arXiv Detail & Related papers (2024-01-13T16:19:49Z) - Unsupervised Learning of Invariance Transformations [105.54048699217668]
We develop an algorithmic framework for finding approximate graph automorphisms.
We discuss how this framework can be used to find approximate automorphisms in weighted graphs in general.
arXiv Detail & Related papers (2023-07-24T17:03:28Z) - Dense Hebbian neural networks: a replica symmetric picture of supervised
learning [4.133728123207142]
We consider dense, associative neural-networks trained by a teacher with supervision.
We investigate their computational capabilities analytically, via statistical-mechanics of spin glasses, and numerically, via Monte Carlo simulations.
arXiv Detail & Related papers (2022-11-25T13:37:47Z) - Measures of Information Reflect Memorization Patterns [53.71420125627608]
We show that the diversity in the activation patterns of different neurons is reflective of model generalization and memorization.
Importantly, we discover that information organization points to the two forms of memorization, even for neural activations computed on unlabelled in-distribution examples.
arXiv Detail & Related papers (2022-10-17T20:15:24Z) - Recurrent neural networks that generalize from examples and optimize by
dreaming [0.0]
We introduce a generalized Hopfield network where pairwise couplings between neurons are built according to Hebb's prescription for on-line learning.
We let the network experience solely a dataset made of a sample of noisy examples for each pattern.
Remarkably, the sleeping mechanisms always significantly reduce the dataset size required to correctly generalize.
arXiv Detail & Related papers (2022-04-17T08:40:54Z) - Data-driven emergence of convolutional structure in neural networks [83.4920717252233]
We show how fully-connected neural networks solving a discrimination task can learn a convolutional structure directly from their inputs.
By carefully designing data models, we show that the emergence of this pattern is triggered by the non-Gaussian, higher-order local structure of the inputs.
arXiv Detail & Related papers (2022-02-01T17:11:13Z) - Dynamic Inference with Neural Interpreters [72.90231306252007]
We present Neural Interpreters, an architecture that factorizes inference in a self-attention network as a system of modules.
inputs to the model are routed through a sequence of functions in a way that is end-to-end learned.
We show that Neural Interpreters perform on par with the vision transformer using fewer parameters, while being transferrable to a new task in a sample efficient manner.
arXiv Detail & Related papers (2021-10-12T23:22:45Z) - PredRNN: A Recurrent Neural Network for Spatiotemporal Predictive
Learning [109.84770951839289]
We present PredRNN, a new recurrent network for learning visual dynamics from historical context.
We show that our approach obtains highly competitive results on three standard datasets.
arXiv Detail & Related papers (2021-03-17T08:28:30Z) - Automatic Recall Machines: Internal Replay, Continual Learning and the
Brain [104.38824285741248]
Replay in neural networks involves training on sequential data with memorized samples, which counteracts forgetting of previous behavior caused by non-stationarity.
We present a method where these auxiliary samples are generated on the fly, given only the model that is being trained for the assessed objective.
Instead the implicit memory of learned samples within the assessed model itself is exploited.
arXiv Detail & Related papers (2020-06-22T15:07:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.