Divergent representations of ethological visual inputs emerge from
supervised, unsupervised, and reinforcement learning
- URL: http://arxiv.org/abs/2112.02027v1
- Date: Fri, 3 Dec 2021 17:18:09 GMT
- Title: Divergent representations of ethological visual inputs emerge from
supervised, unsupervised, and reinforcement learning
- Authors: Grace W. Lindsay, Josh Merel, Tom Mrsic-Flogel, Maneesh Sahani
- Abstract summary: We compare the representations learned by eight different convolutional neural networks.
We find that the network trained with reinforcement learning differs most from the other networks.
- Score: 20.98896935012773
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Artificial neural systems trained using reinforcement, supervised, and
unsupervised learning all acquire internal representations of high dimensional
input. To what extent these representations depend on the different learning
objectives is largely unknown. Here we compare the representations learned by
eight different convolutional neural networks, each with identical ResNet
architectures and trained on the same family of egocentric images, but embedded
within different learning systems. Specifically, the representations are
trained to guide action in a compound reinforcement learning task; to predict
one or a combination of three task-related targets with supervision; or using
one of three different unsupervised objectives. Using representational
similarity analysis, we find that the network trained with reinforcement
learning differs most from the other networks. Through further analysis using
metrics inspired by the neuroscience literature, we find that the model trained
with reinforcement learning has a sparse and high-dimensional representation
wherein individual images are represented with very different patterns of
neural activity. Further analysis suggests these representations may arise in
order to guide long-term behavior and goal-seeking in the RL agent. Our results
provide insights into how the properties of neural representations are
influenced by objective functions and can inform transfer learning approaches.
Related papers
- Identifying Sub-networks in Neural Networks via Functionally Similar Representations [41.028797971427124]
We take a step toward automating the understanding of the network by investigating the existence of distinct sub-networks.
Our approach offers meaningful insights into the behavior of neural networks with minimal human and computational cost.
arXiv Detail & Related papers (2024-10-21T20:19:00Z) - A Probabilistic Model Behind Self-Supervised Learning [53.64989127914936]
In self-supervised learning (SSL), representations are learned via an auxiliary task without annotated labels.
We present a generative latent variable model for self-supervised learning.
We show that several families of discriminative SSL, including contrastive methods, induce a comparable distribution over representations.
arXiv Detail & Related papers (2024-02-02T13:31:17Z) - Predictive auxiliary objectives in deep RL mimic learning in the brain [2.6703221234079946]
We study the effects predictive auxiliary objectives have on representation learning across different modules of a deep reinforcement learning system.
We find that predictive objectives improve and stabilize learning particularly in resource-limited architectures.
We draw a connection between the auxiliary predictive model of the RL system and the hippocampus, an area thought to learn a predictive model to support memory-guided behavior.
arXiv Detail & Related papers (2023-10-09T19:06:25Z) - Evaluating alignment between humans and neural network representations in image-based learning tasks [5.657101730705275]
We tested how well the representations of $86$ pretrained neural network models mapped to human learning trajectories.
We found that while training dataset size was a core determinant of alignment with human choices, contrastive training with multi-modal data (text and imagery) was a common feature of currently publicly available models that predicted human generalisation.
In conclusion, pretrained neural networks can serve to extract representations for cognitive models, as they appear to capture some fundamental aspects of cognition that are transferable across tasks.
arXiv Detail & Related papers (2023-06-15T08:18:29Z) - A picture of the space of typical learnable tasks [12.374133486467013]
We develop information geometric techniques to understand the representations learned by deep networks when they are trained on different tasks.
We shed light on the following phenomena that relate to the structure of the space of tasks.
We use classification tasks constructed from the CIFAR-10 and Imagenet datasets to study these phenomena.
arXiv Detail & Related papers (2022-10-31T01:59:44Z) - Data-driven emergence of convolutional structure in neural networks [83.4920717252233]
We show how fully-connected neural networks solving a discrimination task can learn a convolutional structure directly from their inputs.
By carefully designing data models, we show that the emergence of this pattern is triggered by the non-Gaussian, higher-order local structure of the inputs.
arXiv Detail & Related papers (2022-02-01T17:11:13Z) - Dynamic Inference with Neural Interpreters [72.90231306252007]
We present Neural Interpreters, an architecture that factorizes inference in a self-attention network as a system of modules.
inputs to the model are routed through a sequence of functions in a way that is end-to-end learned.
We show that Neural Interpreters perform on par with the vision transformer using fewer parameters, while being transferrable to a new task in a sample efficient manner.
arXiv Detail & Related papers (2021-10-12T23:22:45Z) - Joint Learning of Neural Transfer and Architecture Adaptation for Image
Recognition [77.95361323613147]
Current state-of-the-art visual recognition systems rely on pretraining a neural network on a large-scale dataset and finetuning the network weights on a smaller dataset.
In this work, we prove that dynamically adapting network architectures tailored for each domain task along with weight finetuning benefits in both efficiency and effectiveness.
Our method can be easily generalized to an unsupervised paradigm by replacing supernet training with self-supervised learning in the source domain tasks and performing linear evaluation in the downstream tasks.
arXiv Detail & Related papers (2021-03-31T08:15:17Z) - Understanding the Role of Individual Units in a Deep Neural Network [85.23117441162772]
We present an analytic framework to systematically identify hidden units within image classification and image generation networks.
First, we analyze a convolutional neural network (CNN) trained on scene classification and discover units that match a diverse set of object concepts.
Second, we use a similar analytic method to analyze a generative adversarial network (GAN) model trained to generate scenes.
arXiv Detail & Related papers (2020-09-10T17:59:10Z) - Self-supervised Learning from a Multi-view Perspective [121.63655399591681]
We show that self-supervised representations can extract task-relevant information and discard task-irrelevant information.
Our theoretical framework paves the way to a larger space of self-supervised learning objective design.
arXiv Detail & Related papers (2020-06-10T00:21:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.