Computational models of learning and synaptic plasticity
- URL: http://arxiv.org/abs/2412.05501v1
- Date: Sat, 07 Dec 2024 02:03:05 GMT
- Title: Computational models of learning and synaptic plasticity
- Authors: Danil Tyulmankov,
- Abstract summary: Many mathematical models of synaptic plasticity have been proposed to explain the diversity of plasticity phenomena observed in biological organisms.
We discuss several fundamental learning paradigms, along with the synaptic plasticity rules that might be used to implement them.
- Score: 1.0878040851638
- License:
- Abstract: Many mathematical models of synaptic plasticity have been proposed to explain the diversity of plasticity phenomena observed in biological organisms. These models range from simple interpretations of Hebb's postulate, which suggests that correlated neural activity leads to increases in synaptic strength, to more complex rules that allow bidirectional synaptic updates, ensure stability, or incorporate additional signals like reward or error. At the same time, a range of learning paradigms can be observed behaviorally, from Pavlovian conditioning to motor learning and memory recall. Although it is difficult to directly link synaptic updates to learning outcomes experimentally, computational models provide a valuable tool for building evidence of this connection. In this chapter, we discuss several fundamental learning paradigms, along with the synaptic plasticity rules that might be used to implement them.
Related papers
- Deep Learning Through A Telescoping Lens: A Simple Model Provides Empirical Insights On Grokking, Gradient Boosting & Beyond [61.18736646013446]
In pursuit of a deeper understanding of its surprising behaviors, we investigate the utility of a simple yet accurate model of a trained neural network.
Across three case studies, we illustrate how it can be applied to derive new empirical insights on a diverse range of prominent phenomena.
arXiv Detail & Related papers (2024-10-31T22:54:34Z) - Artificial Kuramoto Oscillatory Neurons [65.16453738828672]
It has long been known in both neuroscience and AI that ''binding'' between neurons leads to a form of competitive learning.
We introduce Artificial rethinking together with arbitrary connectivity designs such as fully connected convolutional, or attentive mechanisms.
We show that this idea provides performance improvements across a wide spectrum of tasks such as unsupervised object discovery, adversarial robustness, uncertainty, and reasoning.
arXiv Detail & Related papers (2024-10-17T17:47:54Z) - Contrastive Learning in Memristor-based Neuromorphic Systems [55.11642177631929]
Spiking neural networks have become an important family of neuron-based models that sidestep many of the key limitations facing modern-day backpropagation-trained deep networks.
In this work, we design and investigate a proof-of-concept instantiation of contrastive-signal-dependent plasticity (CSDP), a neuromorphic form of forward-forward-based, backpropagation-free learning.
arXiv Detail & Related papers (2024-09-17T04:48:45Z) - Learning with Chemical versus Electrical Synapses -- Does it Make a
Difference? [61.85704286298537]
Bio-inspired neural networks have the potential to advance our understanding of neural computation and improve the state-of-the-art of AI systems.
We conduct experiments with autonomous lane-keeping through a photorealistic autonomous driving simulator to evaluate their performance under diverse conditions.
arXiv Detail & Related papers (2023-11-21T13:07:20Z) - On the Trade-off Between Efficiency and Precision of Neural Abstraction [62.046646433536104]
Neural abstractions have been recently introduced as formal approximations of complex, nonlinear dynamical models.
We employ formal inductive synthesis procedures to generate neural abstractions that result in dynamical models with these semantics.
arXiv Detail & Related papers (2023-07-28T13:22:32Z) - Contrastive-Signal-Dependent Plasticity: Self-Supervised Learning in Spiking Neural Circuits [61.94533459151743]
This work addresses the challenge of designing neurobiologically-motivated schemes for adjusting the synapses of spiking networks.
Our experimental simulations demonstrate a consistent advantage over other biologically-plausible approaches when training recurrent spiking networks.
arXiv Detail & Related papers (2023-03-30T02:40:28Z) - Phenomenological modeling of diverse and heterogeneous synaptic dynamics
at natural density [0.0]
This chapter sheds light on the synaptic organization of the brain from the perspective of computational neuroscience.
It provides an introductory overview on how to account for empirical data in mathematical models, implement them in software, and perform simulations reflecting experiments.
arXiv Detail & Related papers (2022-12-10T19:24:58Z) - Meta-Learning Biologically Plausible Plasticity Rules with Random
Feedback Pathways [0.0]
We develop a novel meta-plasticity approach to discover interpretable, biologically plausible plasticity rules.
Our results highlight the potential of meta-plasticity to discover effective, interpretable learning rules satisfying biological constraints.
arXiv Detail & Related papers (2022-10-28T21:40:56Z) - Spike-based local synaptic plasticity: A survey of computational models
and neuromorphic circuits [1.8464222520424338]
We review historical, bottom-up, and top-down approaches to modeling synaptic plasticity.
We identify computational primitives that can support low-latency and low-power hardware implementations of spike-based learning rules.
arXiv Detail & Related papers (2022-09-30T15:35:04Z) - A Spiking Neuron Synaptic Plasticity Model Optimized for Unsupervised
Learning [0.0]
Spiking neural networks (SNN) are considered as a perspective basis for performing all kinds of learning tasks - unsupervised, supervised and reinforcement learning.
Learning in SNN is implemented through synaptic plasticity - the rules which determine dynamics of synaptic weights depending usually on activity of the pre- and post-synaptic neurons.
arXiv Detail & Related papers (2021-11-12T15:26:52Z) - Embodied Synaptic Plasticity with Online Reinforcement learning [5.6006805285925445]
This paper contributes to bringing the fields of computational neuroscience and robotics closer together by integrating open-source software components from these two fields.
We demonstrate this framework to evaluate Synaptic Plasticity with Online REinforcement learning (SPORE), a reward-learning rule based on synaptic sampling, on two visuomotor tasks.
arXiv Detail & Related papers (2020-03-03T10:29:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.