Self-Supervised Learning Through Efference Copies
- URL: http://arxiv.org/abs/2210.09224v1
- Date: Mon, 17 Oct 2022 16:19:53 GMT
- Title: Self-Supervised Learning Through Efference Copies
- Authors: Franz Scherr, Qinghai Guo, Timoleon Moraitis
- Abstract summary: Self-supervised learning (SSL) methods aim to exploit the abundance of unlabelled data for machine learning (ML)
An SSL framework derived from biological first principles of embodied learning could unify the various SSL methods, help elucidate learning in the brain, and possibly improve ML.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Self-supervised learning (SSL) methods aim to exploit the abundance of
unlabelled data for machine learning (ML), however the underlying principles
are often method-specific. An SSL framework derived from biological first
principles of embodied learning could unify the various SSL methods, help
elucidate learning in the brain, and possibly improve ML. SSL commonly
transforms each training datapoint into a pair of views, uses the knowledge of
this pairing as a positive (i.e. non-contrastive) self-supervisory sign, and
potentially opposes it to unrelated, (i.e. contrastive) negative examples.
Here, we show that this type of self-supervision is an incomplete
implementation of a concept from neuroscience, the Efference Copy (EC).
Specifically, the brain also transforms the environment through efference, i.e.
motor commands, however it sends to itself an EC of the full commands, i.e.
more than a mere SSL sign. In addition, its action representations are likely
egocentric. From such a principled foundation we formally recover and extend
SSL methods such as SimCLR, BYOL, and ReLIC under a common theoretical
framework, i.e. Self-supervision Through Efference Copies (S-TEC). Empirically,
S-TEC restructures meaningfully the within- and between-class representations.
This manifests as improvement in recent strong SSL baselines in image
classification, segmentation, object detection, and in audio. These results
hypothesize a testable positive influence from the brain's motor outputs onto
its sensory representations.
Related papers
- An Empirically Grounded Identifiability Theory Will Accelerate Self-Supervised Learning Research [25.564440860986757]
Self-Supervised Learning (SSL) powers many current AI systems.
Platonic view of SSL suggests that despite different methods and engineering approaches, all representations converge to the same Platonic ideal.
We propose expanding Identifiability Theory (IT) into what we term Singular Identifiability Theory (SITh)
arXiv Detail & Related papers (2025-04-17T17:10:33Z) - Understanding the Role of Equivariance in Self-supervised Learning [51.56331245499712]
equivariant self-supervised learning (E-SSL) learns features to be augmentation-aware.
We identify a critical explaining-away effect in E-SSL that creates a synergy between the equivariant and classification tasks.
We reveal several principles for practical designs of E-SSL.
arXiv Detail & Related papers (2024-11-10T16:09:47Z) - Reverse Engineering Self-Supervised Learning [17.720366509919167]
Self-supervised learning (SSL) is a powerful tool in machine learning.
This paper presents an in-depth empirical analysis of SSL-trained representations.
arXiv Detail & Related papers (2023-05-24T23:15:28Z) - Active Self-Supervised Learning: A Few Low-Cost Relationships Are All
You Need [34.013568381942775]
Self-Supervised Learning (SSL) has emerged as the solution of choice to learn transferable representations from unlabeled data.
In this work, we formalize and generalize this principle through Positive Active Learning (PAL) where an oracle queries semantic relationships between samples.
First, it unveils a theoretically grounded learning framework beyond SSL, based on similarity graphs, that can be extended to tackle supervised and semi-supervised learning depending on the employed oracle.
Second, it provides a consistent algorithm to embed a priori knowledge, e.g. some observed labels, into any SSL losses without any change in the training pipeline.
arXiv Detail & Related papers (2023-03-27T14:44:39Z) - Understanding and Improving the Role of Projection Head in
Self-Supervised Learning [77.59320917894043]
Self-supervised learning (SSL) aims to produce useful feature representations without access to human-labeled data annotations.
Current contrastive learning approaches append a parametrized projection head to the end of some backbone network to optimize the InfoNCE objective.
This raises a fundamental question: Why is a learnable projection head required if we are to discard it after training?
arXiv Detail & Related papers (2022-12-22T05:42:54Z) - On Higher Adversarial Susceptibility of Contrastive Self-Supervised
Learning [104.00264962878956]
Contrastive self-supervised learning (CSL) has managed to match or surpass the performance of supervised learning in image and video classification.
It is still largely unknown if the nature of the representation induced by the two learning paradigms is similar.
We identify the uniform distribution of data representation over a unit hypersphere in the CSL representation space as the key contributor to this phenomenon.
We devise strategies that are simple, yet effective in improving model robustness with CSL training.
arXiv Detail & Related papers (2022-07-22T03:49:50Z) - MC-SSL0.0: Towards Multi-Concept Self-Supervised Learning [26.942174776511237]
Self-supervised pretraining has shown to outperform supervised pretraining for many downstream vision applications.
This superiority is attributed to the negative impact of incomplete labelling of the training images.
This study investigates the possibility of modelling all the concepts present in an image without using labels.
arXiv Detail & Related papers (2021-11-30T12:36:38Z) - Demystifying How Self-Supervised Features Improve Training from Noisy
Labels [16.281091780103736]
We study why and how self-supervised features help networks resist label noise.
Our result shows that, given a quality encoder pre-trained from SSL, a simple linear layer trained by the cross-entropy loss is theoretically robust to symmetric label noise.
arXiv Detail & Related papers (2021-10-18T05:41:57Z) - Self-supervised Learning is More Robust to Dataset Imbalance [65.84339596595383]
We investigate self-supervised learning under dataset imbalance.
Off-the-shelf self-supervised representations are already more robust to class imbalance than supervised representations.
We devise a re-weighted regularization technique that consistently improves the SSL representation quality on imbalanced datasets.
arXiv Detail & Related papers (2021-10-11T06:29:56Z) - LeBenchmark: A Reproducible Framework for Assessing Self-Supervised
Representation Learning from Speech [63.84741259993937]
Self-Supervised Learning (SSL) using huge unlabeled data has been successfully explored for image and natural language processing.
Recent works also investigated SSL from speech.
We propose LeBenchmark: a reproducible framework for assessing SSL from speech.
arXiv Detail & Related papers (2021-04-23T08:27:09Z) - Interventional Few-Shot Learning [88.31112565383457]
We propose a novel Few-Shot Learning paradigm: Interventional Few-Shot Learning.
Code is released at https://github.com/yue-zhongqi/ifsl.
arXiv Detail & Related papers (2020-09-28T01:16:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.