Self-Supervised Learning of Representations for Space Generates
Multi-Modular Grid Cells
- URL: http://arxiv.org/abs/2311.02316v1
- Date: Sat, 4 Nov 2023 03:59:37 GMT
- Title: Self-Supervised Learning of Representations for Space Generates
Multi-Modular Grid Cells
- Authors: Rylan Schaeffer, Mikail Khona, Tzuhsuan Ma, Crist\'obal Eyzaguirre,
Sanmi Koyejo, Ila Rani Fiete
- Abstract summary: mammalian lineage has developed striking spatial representations.
One important spatial representation is the Nobel-prize winning grid cells.
Nobel-prize winning grid cells represent self-location, a local and aperiodic quantity.
- Score: 16.208253624969142
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: To solve the spatial problems of mapping, localization and navigation, the
mammalian lineage has developed striking spatial representations. One important
spatial representation is the Nobel-prize winning grid cells: neurons that
represent self-location, a local and aperiodic quantity, with seemingly bizarre
non-local and spatially periodic activity patterns of a few discrete periods.
Why has the mammalian lineage learnt this peculiar grid representation?
Mathematical analysis suggests that this multi-periodic representation has
excellent properties as an algebraic code with high capacity and intrinsic
error-correction, but to date, there is no satisfactory synthesis of core
principles that lead to multi-modular grid cells in deep recurrent neural
networks. In this work, we begin by identifying key insights from four families
of approaches to answering the grid cell question: coding theory, dynamical
systems, function optimization and supervised deep learning. We then leverage
our insights to propose a new approach that combines the strengths of all four
approaches. Our approach is a self-supervised learning (SSL) framework -
including data, data augmentations, loss functions and a network architecture -
motivated from a normative perspective, without access to supervised position
information or engineering of particular readout representations as needed in
previous approaches. We show that multiple grid cell modules can emerge in
networks trained on our SSL framework and that the networks and emergent
representations generalize well outside their training distribution. This work
contains insights for neuroscientists interested in the origins of grid cells
as well as machine learning researchers interested in novel SSL frameworks.
Related papers
- The Dynamic Net Architecture: Learning Robust and Holistic Visual Representations Through Self-Organizing Networks [3.9848584845601014]
We present a novel intelligent-system architecture called "Dynamic Net Architecture" (DNA)
DNA relies on recurrence-stabilized networks and discuss it in application to vision.
arXiv Detail & Related papers (2024-07-08T06:22:10Z) - Coding schemes in neural networks learning classification tasks [52.22978725954347]
We investigate fully-connected, wide neural networks learning classification tasks.
We show that the networks acquire strong, data-dependent features.
Surprisingly, the nature of the internal representations depends crucially on the neuronal nonlinearity.
arXiv Detail & Related papers (2024-06-24T14:50:05Z) - Multiway Multislice PHATE: Visualizing Hidden Dynamics of RNNs through Training [6.326396282553267]
Recurrent neural networks (RNNs) are a widely used tool for sequential data analysis, however, they are still often seen as black boxes of computation.
Here, we present Multiway Multislice PHATE (MM-PHATE), a novel method for visualizing the evolution of RNNs' hidden states.
arXiv Detail & Related papers (2024-06-04T05:05:27Z) - Riemannian Residual Neural Networks [58.925132597945634]
We show how to extend the residual neural network (ResNet)
ResNets have become ubiquitous in machine learning due to their beneficial learning properties, excellent empirical results, and easy-to-incorporate nature when building varied neural networks.
arXiv Detail & Related papers (2023-10-16T02:12:32Z) - How neural networks learn to classify chaotic time series [77.34726150561087]
We study the inner workings of neural networks trained to classify regular-versus-chaotic time series.
We find that the relation between input periodicity and activation periodicity is key for the performance of LKCNN models.
arXiv Detail & Related papers (2023-06-04T08:53:27Z) - Data-driven emergence of convolutional structure in neural networks [83.4920717252233]
We show how fully-connected neural networks solving a discrimination task can learn a convolutional structure directly from their inputs.
By carefully designing data models, we show that the emergence of this pattern is triggered by the non-Gaussian, higher-order local structure of the inputs.
arXiv Detail & Related papers (2022-02-01T17:11:13Z) - A neural anisotropic view of underspecification in deep learning [60.119023683371736]
We show that the way neural networks handle the underspecification of problems is highly dependent on the data representation.
Our results highlight that understanding the architectural inductive bias in deep learning is fundamental to address the fairness, robustness, and generalization of these systems.
arXiv Detail & Related papers (2021-04-29T14:31:09Z) - PredRNN: A Recurrent Neural Network for Spatiotemporal Predictive
Learning [109.84770951839289]
We present PredRNN, a new recurrent network for learning visual dynamics from historical context.
We show that our approach obtains highly competitive results on three standard datasets.
arXiv Detail & Related papers (2021-03-17T08:28:30Z) - Supervised Learning with First-to-Spike Decoding in Multilayer Spiking
Neural Networks [0.0]
We propose a new supervised learning method that can train multilayer spiking neural networks to solve classification problems.
The proposed learning rule supports multiple spikes fired by hidden neurons, and yet is stable by relying on firstspike responses generated by a deterministic output layer.
We also explore several distinct spike-based encoding strategies in order to form compact representations of input data.
arXiv Detail & Related papers (2020-08-16T15:34:48Z) - Grid Cells Are Ubiquitous in Neural Networks [0.0]
Grid cells are believed to play an important role in both spatial and non-spatial cognition tasks.
Recent study observed the emergence of grid cells in an LSTM for path integration.
arXiv Detail & Related papers (2020-03-07T01:40:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.