Equivariant Contrastive Learning
- URL: http://arxiv.org/abs/2111.00899v1
- Date: Thu, 28 Oct 2021 17:21:33 GMT
- Title: Equivariant Contrastive Learning
- Authors: Rumen Dangovski, Li Jing, Charlotte Loh, Seungwook Han, Akash
Srivastava, Brian Cheung, Pulkit Agrawal and Marin Solja\v{c}i\'c
- Abstract summary: In state-of-the-art self-supervised learning (SSL) pre-training produces semantically good representations.
We extend popular SSL methods to a more general framework which we name Equivariant Self-Supervised Learning (E-SSL)
We demonstrate E-SSL's effectiveness empirically on several popular computer vision benchmarks.
- Score: 20.369942206674576
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In state-of-the-art self-supervised learning (SSL) pre-training produces
semantically good representations by encouraging them to be invariant under
meaningful transformations prescribed from human knowledge. In fact, the
property of invariance is a trivial instance of a broader class called
equivariance, which can be intuitively understood as the property that
representations transform according to the way the inputs transform. Here, we
show that rather than using only invariance, pre-training that encourages
non-trivial equivariance to some transformations, while maintaining invariance
to other transformations, can be used to improve the semantic quality of
representations. Specifically, we extend popular SSL methods to a more general
framework which we name Equivariant Self-Supervised Learning (E-SSL). In E-SSL,
a simple additional pre-training objective encourages equivariance by
predicting the transformations applied to the input. We demonstrate E-SSL's
effectiveness empirically on several popular computer vision benchmarks.
Furthermore, we demonstrate usefulness of E-SSL for applications beyond
computer vision; in particular, we show its utility on regression problems in
photonics science. We will release our code.
Related papers
- Understanding the Role of Equivariance in Self-supervised Learning [51.56331245499712]
equivariant self-supervised learning (E-SSL) learns features to be augmentation-aware.
We identify a critical explaining-away effect in E-SSL that creates a synergy between the equivariant and classification tasks.
We reveal several principles for practical designs of E-SSL.
arXiv Detail & Related papers (2024-11-10T16:09:47Z) - Equivariant Spatio-Temporal Self-Supervision for LiDAR Object Detection [37.142470149311904]
We propose atemporal equivariant learning framework by considering both spatial and temporal augmentations jointly.
We show our pre-training method for 3D object detection which outperforms existing equivariant and invariant approaches in many settings.
arXiv Detail & Related papers (2024-04-17T20:41:49Z) - ESCL: Equivariant Self-Contrastive Learning for Sentence Representations [16.601370864663213]
We propose an Equivariant Self-Contrastive Learning (ESCL) method to make full use of sensitive transformations.
The proposed method achieves better results while using fewer learning parameters compared to previous methods.
arXiv Detail & Related papers (2023-03-09T09:52:28Z) - The Lie Derivative for Measuring Learned Equivariance [84.29366874540217]
We study the equivariance properties of hundreds of pretrained models, spanning CNNs, transformers, and Mixer architectures.
We find that many violations of equivariance can be linked to spatial aliasing in ubiquitous network layers, such as pointwise non-linearities.
For example, transformers can be more equivariant than convolutional neural networks after training.
arXiv Detail & Related papers (2022-10-06T15:20:55Z) - Learning Instance-Specific Augmentations by Capturing Local Invariances [62.70897571389785]
InstaAug is a method for automatically learning input-specific augmentations from data.
We empirically demonstrate that InstaAug learns meaningful input-dependent augmentations for a wide range of transformation classes.
arXiv Detail & Related papers (2022-05-31T18:38:06Z) - A Generic Self-Supervised Framework of Learning Invariant Discriminative
Features [9.614694312155798]
This paper proposes a generic SSL framework based on a constrained self-labelling assignment process.
The proposed training strategy outperforms a majority of state-of-the-art representation learning methods based on AE structures.
arXiv Detail & Related papers (2022-02-14T18:09:43Z) - Exploring Complementary Strengths of Invariant and Equivariant
Representations for Few-Shot Learning [96.75889543560497]
In many real-world problems, collecting a large number of labeled samples is infeasible.
Few-shot learning is the dominant approach to address this issue, where the objective is to quickly adapt to novel categories in presence of a limited number of samples.
We propose a novel training mechanism that simultaneously enforces equivariance and invariance to a general set of geometric transformations.
arXiv Detail & Related papers (2021-03-01T21:14:33Z) - Learning Invariances in Neural Networks [51.20867785006147]
We show how to parameterize a distribution over augmentations and optimize the training loss simultaneously with respect to the network parameters and augmentation parameters.
We can recover the correct set and extent of invariances on image classification, regression, segmentation, and molecular property prediction from a large space of augmentations.
arXiv Detail & Related papers (2020-10-22T17:18:48Z) - Plannable Approximations to MDP Homomorphisms: Equivariance under
Actions [72.30921397899684]
We introduce a contrastive loss function that enforces action equivariance on the learned representations.
We prove that when our loss is zero, we have a homomorphism of a deterministic Markov Decision Process.
We show experimentally that for deterministic MDPs, the optimal policy in the abstract MDP can be successfully lifted to the original MDP.
arXiv Detail & Related papers (2020-02-27T08:29:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.