Towards Democratizing Joint-Embedding Self-Supervised Learning
- URL: http://arxiv.org/abs/2303.01986v1
- Date: Fri, 3 Mar 2023 14:55:44 GMT
- Title: Towards Democratizing Joint-Embedding Self-Supervised Learning
- Authors: Florian Bordes, Randall Balestriero, Pascal Vincent
- Abstract summary: We show that it is possible to train SimCLR to learn useful representations, while using a single image patch as negative example.
In the hope to democratize JE-SSL, we introduce an optimized PyTorch library for SSL.
- Score: 17.59181163979478
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Joint Embedding Self-Supervised Learning (JE-SSL) has seen rapid developments
in recent years, due to its promise to effectively leverage large unlabeled
data. The development of JE-SSL methods was driven primarily by the search for
ever increasing downstream classification accuracies, using huge computational
resources, and typically built upon insights and intuitions inherited from a
close parent JE-SSL method. This has led unwittingly to numerous pre-conceived
ideas that carried over across methods e.g. that SimCLR requires very large
mini batches to yield competitive accuracies; that strong and computationally
slow data augmentations are required. In this work, we debunk several such
ill-formed a priori ideas in the hope to unleash the full potential of JE-SSL
free of unnecessary limitations. In fact, when carefully evaluating
performances across different downstream tasks and properly optimizing
hyper-parameters of the methods, we most often -- if not always -- see that
these widespread misconceptions do not hold. For example we show that it is
possible to train SimCLR to learn useful representations, while using a single
image patch as negative example, and simple Gaussian noise as the only data
augmentation for the positive pair. Along these lines, in the hope to
democratize JE-SSL and to allow researchers to easily make more extensive
evaluations of their methods, we introduce an optimized PyTorch library for
SSL.
Related papers
- Reinforcement Learning-Guided Semi-Supervised Learning [20.599506122857328]
We propose a novel Reinforcement Learning Guided SSL method, RLGSSL, that formulates SSL as a one-armed bandit problem.
RLGSSL incorporates a carefully designed reward function that balances the use of labeled and unlabeled data to enhance generalization performance.
We demonstrate the effectiveness of RLGSSL through extensive experiments on several benchmark datasets and show that our approach achieves consistent superior performance compared to state-of-the-art SSL methods.
arXiv Detail & Related papers (2024-05-02T21:52:24Z) - BECLR: Batch Enhanced Contrastive Few-Shot Learning [1.450405446885067]
Unsupervised few-shot learning aspires to bridge this gap by discarding the reliance on annotations at training time.
We propose a novel Dynamic Clustered mEmory (DyCE) module to promote a highly separable latent representation space.
We then tackle the, somehow overlooked yet critical, issue of sample bias at the few-shot inference stage.
arXiv Detail & Related papers (2024-02-04T10:52:43Z) - Semi-Supervised Class-Agnostic Motion Prediction with Pseudo Label
Regeneration and BEVMix [59.55173022987071]
We study the potential of semi-supervised learning for class-agnostic motion prediction.
Our framework adopts a consistency-based self-training paradigm, enabling the model to learn from unlabeled data.
Our method exhibits comparable performance to weakly and some fully supervised methods.
arXiv Detail & Related papers (2023-12-13T09:32:50Z) - Progressive Feature Adjustment for Semi-supervised Learning from
Pretrained Models [39.42802115580677]
Semi-supervised learning (SSL) can leverage both labeled and unlabeled data to build a predictive model.
Recent literature suggests that naively applying state-of-the-art SSL with a pretrained model fails to unleash the full potential of training data.
We propose to use pseudo-labels from the unlabelled data to update the feature extractor that is less sensitive to incorrect labels.
arXiv Detail & Related papers (2023-09-09T01:57:14Z) - Data-Efficient Contrastive Self-supervised Learning: Most Beneficial
Examples for Supervised Learning Contribute the Least [14.516008359896421]
Self-supervised learning (SSL) learns high-quality representations from large pools of unlabeled training data.
As datasets grow larger, it becomes crucial to identify the examples that contribute the most to learning such representations.
We prove that examples that contribute the most to contrastive SSL are those that have the most similar augmentations to other examples.
arXiv Detail & Related papers (2023-02-18T00:15:06Z) - Benchmark for Uncertainty & Robustness in Self-Supervised Learning [0.0]
Self-Supervised Learning is crucial for real-world applications, especially in data-hungry domains such as healthcare and self-driving cars.
In this paper, we explore variants of SSL methods, including Jigsaw Puzzles, Context, Rotation, Geometric Transformations Prediction for vision, as well as BERT and GPT for language tasks.
Our goal is to create a benchmark with outputs from experiments, providing a starting point for new SSL methods in Reliable Machine Learning.
arXiv Detail & Related papers (2022-12-23T15:46:23Z) - MaxMatch: Semi-Supervised Learning with Worst-Case Consistency [149.03760479533855]
We propose a worst-case consistency regularization technique for semi-supervised learning (SSL)
We present a generalization bound for SSL consisting of the empirical loss terms observed on labeled and unlabeled training data separately.
Motivated by this bound, we derive an SSL objective that minimizes the largest inconsistency between an original unlabeled sample and its multiple augmented variants.
arXiv Detail & Related papers (2022-09-26T12:04:49Z) - Improving Self-Supervised Learning by Characterizing Idealized
Representations [155.1457170539049]
We prove necessary and sufficient conditions for any task invariant to given data augmentations.
For contrastive learning, our framework prescribes simple but significant improvements to previous methods.
For non-contrastive learning, we use our framework to derive a simple and novel objective.
arXiv Detail & Related papers (2022-09-13T18:01:03Z) - OpenLDN: Learning to Discover Novel Classes for Open-World
Semi-Supervised Learning [110.40285771431687]
Semi-supervised learning (SSL) is one of the dominant approaches to address the annotation bottleneck of supervised learning.
Recent SSL methods can effectively leverage a large repository of unlabeled data to improve performance while relying on a small set of labeled data.
This work introduces OpenLDN that utilizes a pairwise similarity loss to discover novel classes.
arXiv Detail & Related papers (2022-07-05T18:51:05Z) - Self-supervised Learning is More Robust to Dataset Imbalance [65.84339596595383]
We investigate self-supervised learning under dataset imbalance.
Off-the-shelf self-supervised representations are already more robust to class imbalance than supervised representations.
We devise a re-weighted regularization technique that consistently improves the SSL representation quality on imbalanced datasets.
arXiv Detail & Related papers (2021-10-11T06:29:56Z) - Understanding self-supervised Learning Dynamics without Contrastive
Pairs [72.1743263777693]
Contrastive approaches to self-supervised learning (SSL) learn representations by minimizing the distance between two augmented views of the same data point.
BYOL and SimSiam, show remarkable performance it without negative pairs.
We study the nonlinear learning dynamics of non-contrastive SSL in simple linear networks.
arXiv Detail & Related papers (2021-02-12T22:57:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.